MLPerf Inference 5.0 signals the rise of large language models, with LLAMA 2 70B surpassing ResNet-50 in submissions and driving next-gen AI performance across compute platforms.
MemryX, a provider of edge AI acceleration hardware, recently closed its latest round of funding, serving as a potential bellwether for the next growth edge in AI compute.
From VAST Data to Weka, Graid to Solidigm — storage disruptors shined bright at NVIDIA GTC 2025. Here’s how storage innovators are redefining AI infrastructure and why it matters to the future of AI.
Deloitte and VAST Data share how secure data pipelines and system-level integration are supporting the shift to scalable, agentic AI across enterprise environments.
This video explores how Nebius and VAST Data are partnering to power enterprise AI with full-stack cloud infrastructure—spanning compute, storage, and data services for training and inference at scale.
Weka’s new memory grid raises new questions about AI data architecture—exploring how shifts in interface speeds and memory tiers may reshape performance, scale, and deployment strategies.
In 2025, the internet’s fragility and AI’s complexity collided in public. The big vendors responded by buying the pieces they need to sell the integrated story that they have AI risk under control.
Robots aren’t going to fold your laundry by February. But Voice of Innovation Niv Sundaram predicts that an urgent caregiver shortage will move humanoids “from warehouses to living rooms” in 2026.
Marvell is inking a deal for optical interconnect startup Celestial AI in a massive bet that the industry has shifted from being compute-constrained to bandwidth-constrained.
As AI training pushes data centers to unprecedented power densities, researchers reveal an affordable solution that lets computing thrive on fluctuating renewable energy.
In Part 1 of Voice of Innovation Matty Bakkeren’s 2026 predictions series, he explores how AI, power, cooling, and supply chains are reshaping data center infrastructure for a utility-scale future.
As up to 10 million jobs disappear and quality content moves behind paywalls, the question isn’t if AI will reshape society. It’s whether 2026 is the year we’ll control the burn or watch it spread.
Runpod head of engineering Brennen Smith joins a Data Insights episode to unpack GPU-dense clouds, hidden storage bottlenecks, and a “universal orchestrator” for long-running AI agents at scale.
From CPU orchestration to scaling efficiency in networks, leaders reveal how to assess your use case, leverage existing infrastructure, and productize AI instead of just experimenting.
From the OCP Global Summit, hear why 50% GPU utilization is a “civilization-level” problem, and why open standards are key to unlocking underutilized compute capacity.
In the Arena: Allyson Klein with Axelera CMO Alexis Crowell on inference-first AI silicon, a customer-driven SDK, and what recent tapeouts reveal about the roadmap.
In this episode of Data Insights, host Allyson Klein and co-host Jeniece Wnorowski sit down with Dr. Rohith Vangalla of Optum to discuss the future of AI in healthcare.
From OCP Summit, Metrum AI CEO Steen Graham unpacks multi-agent infrastructure, SSD-accelerated RAG, and the memory-to-storage shift—plus a 2026 roadmap to boost GPU utilization, uptime, and time-to-value.
In this podcast, MLCommons President Peter Mattson discusses their just-released AILuminate benchmark, AI safety, and how global collaboration is driving trust and innovation in AI deployment.
In this episode, Eric Kavanagh anticipates AI's evolving role in enterprise for 2025. He explores practical applications, the challenges of generative AI, future advancements in co-pilots and agents, and more.
Peter Dueben of European Centre for Medium-Range Weather Forecasts explores the role of HPC and AI in advancing weather modeling, tackling climate challenges, and scaling predictions to the kilometer level.
David Kanter discusses MLCommons' role in setting benchmarks for AI performance, fostering industry-wide collaboration, and driving advancements in machine learning capabilities.
Join Allyson Klein and Jeniece Wnorowski in this episode of Data Insights as they discuss key takeaways from the 2024 OCP Summit with Scott Shadley, focusing on AI advancements and storage innovations.
In this episode of Data Insights by Solidigm, Ravi Kuppuswamy of AMD unpacks the company’s innovations in data center computing and how they adapt to AI demands while supporting traditional workloads.