UPDATED 16:00 EST / NOVEMBER 18 2025

Isaiah Weiner, head of product management, core software, at WekaIO, and Avi Shetty, senior director of AI enablement and partnerships at Solidigm, talk to theCUBE about storage efficiency at SC25. AI

Storage efficiency takes center stage as AI scale reshapes the data pipeline

Storage efficiency is becoming a practical requirement for keeping performance, density and cost in check as organizations build out larger, faster data systems.

Rising demands on compute and networking are putting new pressure on the storage layer, making speed and responsiveness essential across modern deployments. That shift is showing up clearly in customer conversations, according to Avi Shetty (pictured, right), senior director of AI enablement and partnerships at Solidigm, a trademark of SK Hynix NAND Products Solutions Corp.

“I think last year when I spoke to you guys, GPU and [High Bandwidth Memory] were kind of the favorite childs,” Shetty said. “Over the last year, we’ve seen storage kind of put itself in its place where you’ve seen usages and certain solutions which have exposed the need for having high performing, reliant, scalable, high-dense storage solutions. That’s exactly what we’ve been doing at Solidigm.”

Shetty and Isaiah Weiner (left), head of product management, core software, at WekaIO Inc., spoke with theCUBE’s Dave Vellante and Savannah Peterson at SC25, during an exclusive broadcast on theCUBE, SiliconANGLE Media’s livestreaming studio. They explored how storage efficiency is becoming a central force in scaling modern AI workloads across training, inference and high-performance infrastructure. (* Disclosure below.)

Storage efficiency shapes next-generation AI design

As AI workloads expand across training clusters, inference endpoints and distributed data pipelines, storage efficiency is becoming a defining architectural requirement. Organizations want predictable responsiveness, even as GPU cores multiply and PCIe and networking systems ramp up bandwidth. The trend is visible across deployments where balanced systems outperform raw compute alone, Shetty noted.

“Two trends have emerged,” he said. “When it comes to storage, you need low latency, fast super performing storage. On the other end you need high-dense cost-effective storage solutions. A lot [is] happening in the storage world for us at Solidigm. We are excited to work with Weka to bring all of this storage goodness to our customers in an easy, scalable solution way.”

This new generation of SSD design is directly influencing how AI systems scale. Higher-performance tiers feed GPUs with precision, while high-density tiers support global inference footprints where cost-per-bit and longevity matter as much as throughput. Both are becoming essential as customers balance power constraints, model iteration cycles and long-term operating costs, according to Weiner.

“We’ve been working a long time together,” he said, referring to Weka’s partnership with Solidigm. “They were the first QLC drives that we actually had in our hands. This morning, we announced improvements for Weka customers that they can get 90% with QLC, they can get 90% of the performance, but at a massive reduction.”

Inference at scale is sharpening these priorities further. User-facing applications demand seamless responsiveness, forcing organizations to rethink how fast models load, refresh and transition between requests. The economics of token generation and GPU utilization are pushing companies toward storage-centric designs that minimize idle cycles and maximize throughput across memory and flash layers.

“Inference is what makes AI real,” Peterson said. “None of the stuff that we’re talking about in terms of its real-world application and experience exists if inference isn’t optimized.”

Here’s the complete video interview, part of SiliconANGLE’s and theCUBE’s coverage of SC25:

(* Disclosure: Solidigm sponsored this segment of theCUBE. Neither Solidigm nor other sponsors have editorial control over content on theCUBE or SiliconANGLE.)

Photo: SiliconANGLE

A message from John Furrier, co-founder of SiliconANGLE:

Support our mission to keep content open and free by engaging with theCUBE community. Join theCUBE’s Alumni Trust Network, where technology leaders connect, share intelligence and create opportunities.

  • 15M+ viewers of theCUBE videos, powering conversations across AI, cloud, cybersecurity and more
  • 11.4k+ theCUBE alumni — Connect with more than 11,400 tech and business leaders shaping the future through a unique trusted-based network.
About SiliconANGLE Media
SiliconANGLE Media is a recognized leader in digital media innovation, uniting breakthrough technology, strategic insights and real-time audience engagement. As the parent company of SiliconANGLE, theCUBE Network, theCUBE Research, CUBE365, theCUBE AI and theCUBE SuperStudios — with flagship locations in Silicon Valley and the New York Stock Exchange — SiliconANGLE Media operates at the intersection of media, technology and AI.

Founded by tech visionaries John Furrier and Dave Vellante, SiliconANGLE Media has built a dynamic ecosystem of industry-leading digital media brands that reach 15+ million elite tech professionals. Our new proprietary theCUBE AI Video Cloud is breaking ground in audience interaction, leveraging theCUBEai.com neural network to help technology companies make data-driven decisions and stay at the forefront of industry conversations.