NVIDIA and SK hynix Begin Joint Development of AI-Focused SSDs

NVIDIA and SK hynix have expanded their AI hardware collaboration beyond high-bandwidth memory, moving into joint development of AI-focused solid-state drives designed specifically for inference workloads.

According to multiple industry reports, the companies are working on storage technology intended to remove performance bottlenecks that GPUs and HBM alone can no longer solve. The effort reflects a broader shift in how AI systems are being designed as inference workloads continue to scale.

The project is internally referred to as “Storage Next” at NVIDIA and “AI-N P” at SK hynix. The long-term goal, according to the reports, is to achieve ten times the performance of existing enterprise SSDs, targeting 100 million IOPS by 2027.

Storage Emerging as a Bottleneck in AI Inference

While recent advancements in AI hardware have primarily focused on GPUs and high-bandwidth memory, inference workloads place unique demands on system architecture. According to Chosun Biz, large language models require constant access to massive parameter sets, user context, and external data sources that often exceed the capacity of on-package memory.

SK hynix has revealed that advanced inference systems may require several terabytes of accessible data, forcing operators to rely on multi-GPU configurations. This increases latency, power consumption, and overall system costs, making storage performance a limiting factor in large-scale deployments.

AI SSDs Positioned as a New System Layer

According to Biztalktalk, SK hynix is ​​designing the upcoming AI SSDs as a pseudo-memory layer that sits between system DRAM and long-term storage. Unlike traditional enterprise SSDs, these drives are being optimized at both the NAND architecture and controller levels to support the high-frequency data access patterns common in inference workloads.

The goal is to reduce GPU idle time by improving data availability rather than simply increasing raw compute power. Initial proof-of-concept testing is already underway, and a working prototype is expected during the next development phase.

Parallel Work on High-Bandwidth NAND with SanDisk

In parallel with its work with NVIDIA, SK hynix is ​​also collaborating with SanDisk on high-bandwidth flash, a technology that applies a stacking concept similar to HBM while retaining the non-volatile characteristics of NAND. According to Chosun Biz, an alpha version of the high-bandwidth flash is expected next year, with customer evaluations planned for 2027.

Market projections cited in the same reports suggest that AI-focused NAND solutions could grow rapidly as inference workloads increase in cloud and enterprise environments.

Comparison diagram showing traditional SSD data flow versus AI SSD with direct GPU processing
Diagram illustrating how AI-focused SSDs process data directly with GPUs, reducing CPU bottlenecks compared to traditional SSDs.

Concerns Regarding Impact on NAND Supply

According to Wccftech, the shift towards storage designed for AI is already raising concerns in the semiconductor supply chain. Analysts warn that widespread adoption of AI SSDs could strain NAND flash production, similar to the supply shortages currently affecting DRAM and HBM.

Cloud service providers and AI companies are already consuming large quantities of memory and storage components, which could lead to reduced availability for traditional enterprise and consumer markets if capacity doesn’t increase accordingly.

Also Read: Best SSD Laptops

What This Means for AI Infrastructure

Overall, the reports indicate a significant shift in AI system design. Instead of relying solely on powerful GPUs and large memory pools, vendors are now viewing storage as a performance-critical component, especially for inference and personalized AI services.

If the performance targets outlined by NVIDIA and SK hynix are achieved, AI-optimized SSDs could reduce inference latency, lower operating costs, and make it easier to deploy large models at scale.

Sources:
Chosun Biz
Wccftech

Related Articles

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Latest Articles