The race to trillion-parameter model training in AI is on, and this company thinks it can manage it for less than $100,000

Collapse
X
 
  • Time
  • Show
Clear All
new posts
  • PCHF IT Feeds
    PCHF Bot
    • Jan 2015
    • 54573

    #1

    The race to trillion-parameter model training in AI is on, and this company thinks it can manage it for less than $100,000



    [ul]
    [li]Phison’s SSD strategy slashes AI training costs from $3 million to $100,000[/li][li]aiDAPTIV+ software shifts AI workloads from GPUs to SSDs efficiently[/li][li]SSDs could replace costly GPUs in massive AI model training[/li][/ul]



    The development of AI models has become increasingly costly as their size and complexity grow, requiring massive computational resources with GPUs playing a central role in handling the workload.

    Phison, a key player in portable SSDs, has unveiled a new solution that aims to drastically reduce the cost of training a 1 trillion parameter model by shifting some of the processing load from GPUs to SSDs, bringing the estimated $3 million operational expense down to just $100,000.

    Phison’s strategy involves integrating its aiDAPTIV+ software with high-performance SSDs to handle some AI tool processing tasks traditionally managed by GPUs while also incorporating NVIDIA’s GH200 Superchip to enhance performance and keep costs manageable.

    [HEADING=1]AI model growth and the trillion-parameter milestone[/HEADING]

    Phison expects the AI industry to reach the 1 trillion parameter milestone before 2026.

    According to the company, model sizes have expanded rapidly, moving from 69 billion parameters in Llama 2 (2023) to 405 billion with Llama 3.1 (2024), followed by DeepSeek R3’s 671 billion parameters (2025).

    If this pattern continues, a trillion-parameter model could be unveiled before the end of 2025, marking a significant leap in AI capabilities.

    In addition, it believes that its solution can significantly reduce the number of GPUs needed to run large-scale AI models by shifting some of the processing tasks away from GPUs to the largest SSDs and this approach could bring down training costs to just 3% of current projections (97% savings), or less than 1/25 of the usual operating expenses.

    Phison has already collaborated with Maingear to launch AI workstations powered by Intel Xeon W7-3455 CPUs, signaling its commitment to reshaping AI hardware.

    As companies seek cost-effective ways to train massive AI models, innovations in SSD technology could play a crucial role in driving efficiency gains while external HDD options remain relevant for long-term data storage.

    The push for cheaper AI training solutions gained momentum after DeepSeek made headlines earlier this year when its DeepSeek R1 model demonstrated that cutting-edge AI could be developed at a fraction of the usual cost, with 95% fewer chips and reportedly requiring only $6 million for training.

    Via Tweaktown

    [HEADING=2]You may also like[/HEADING]
    [ul]
    [li]These are the best AI website builders around[/li][li]Take a look at the best firewalls on offer right now[/li][li]Seagate teams with Nvidia to build an NVMe hard drive proof of concept, more than 3 years after its last effort[/li][/ul]

    Continue reading…
Working...