Jim Fister dives deep into the intricacies of system memory, latency, and data management, exploring how modern computing architectures handle data retrieval and processing.
AI is driving an acceleration of compute demands fueled by the proliferation of large language models across industry use cases. This requirement comes as traditional semiconductor technology pushes against the laws of physics with the slowing of Moore’s Law.
This week, Contextual AI partnered with WEKA to deliver enterprise AI services on Google Cloud using RAG 2.0 from Facebook AI Research. WEKA's platform boosted performance, achieving a 3X increase in key AI use cases, 4X faster model checkpointing, and reduced costs.
Data center industry veteran Lynn Comp shares insights from her career, discussing the importance of adaptability, emotional IQ, and practical tech. She highlights three patterns: customers prefer reliability over high-performance solutions, network limitations can negate compute power, and economic realities can override technological enthusiasm.
The upcoming OCP Summit in October will feature nineteen top-tier sponsors, up from just three in previous years, highlighting OCP’s role in AI infrastructure innovation. TechArena is excited to be a media sponsor, covering sustainable infrastructure, Sonic’s future, memory advancements, and power and cooling solutions with daily podcasts, video interviews, and stories.
In this blog post, industry veteran Jim Fister explores the evolution of data centers from early 2000s servers to modern AI/ML racks. Highlighting engineering and logistical challenges, he emphasizes the need for ongoing innovation and celebrates engineers while anticipating future advancements to meet increasing power demands.
LLMs have given attackers new angles. Fortinet showed, step by step, how AI-driven probes escalate—and how FortiGate, FortiWeb, FortiAnalyzer, and FortiSOAR close the door without slowing the business.
From racks and liquid loops to data placement and standards pace, five takeaways from CoreWeave, Dell, NVIDIA, Solidigm, and VAST Data on building AI factories that keep accelerators busy and dollars well-spent.
At Cloud Field Day 24, Oxide outlines a vertically integrated rack—custom hypervisor, integrated power/network, and open integrations—aimed at bringing hyperscale efficiency and faster deploys to enterprise DCs.
Presenting at Cloud Field Day 24, Pure pitched fleet-level automation across mixed environments as the antidote to storage silos, promising one control plane for legacy systems and modern workloads.
A 2025 field guide for architects: why Arm’s software gravity and hyperscaler adoption make it the low-friction path today, where RISC-V is gaining ground, and the curveballs that could reshape both.
The cloud security architect who came from networking explains her framework for separating tech hype from genuine innovation, and why stepping away is key to solving hard problems.
Allyson Klein and co-host Jeniece Wnorowski sit down with Arm’s Eddie Ramirez to unpack Arm Total Design’s growth, the FCSA chiplet spec contribution to OCP, a new board seat, and how storage fits AI’s surge.
Midas Immersion Cooling CEO Scott Sickmiller joins a Data Insights episode at OCP 2025 to demystify single-phase immersion, natural vs. forced convection, and what it takes to do liquid cooling at AI scale.
From hyperscale direct-to-chip to micron-level realities: Darren Burgess (Castrol) explains dielectric fluids, additive packs, particle risks, and how OCP standards keep large deployments on track.
From OCP San Jose, PEAK:AIO’s Roger Cummings explains how workload-aware file systems, richer memory tiers, and capturing intelligence at the edge reduce cost and complexity.
Innovative power delivery unlocks a shift in data-center design. CelLink PowerPlane routes thousands of amps in a flat, flexible circuit—cutting cabling and accelerating AI factory builds.
Recorded live at OCP in San Jose, Allyson Klein talks with CESQ’s Lesya Dymyd about hybrid quantum-classical computing, the new Maison du Quantique, and how real-world use cases may emerge over the next 5–7 years.
Allyson Klein and co-host Jeniece Wnorowski sit down with Arm’s Eddie Ramirez to unpack Arm Total Design’s growth, the FCSA chiplet spec contribution to OCP, a new board seat, and how storage fits AI’s surge.
Midas Immersion Cooling CEO Scott Sickmiller joins a Data Insights episode at OCP 2025 to demystify single-phase immersion, natural vs. forced convection, and what it takes to do liquid cooling at AI scale.
From hyperscale direct-to-chip to micron-level realities: Darren Burgess (Castrol) explains dielectric fluids, additive packs, particle risks, and how OCP standards keep large deployments on track.
From OCP San Jose, PEAK:AIO’s Roger Cummings explains how workload-aware file systems, richer memory tiers, and capturing intelligence at the edge reduce cost and complexity.
Innovative power delivery unlocks a shift in data-center design. CelLink PowerPlane routes thousands of amps in a flat, flexible circuit—cutting cabling and accelerating AI factory builds.
Recorded live at OCP in San Jose, Allyson Klein talks with CESQ’s Lesya Dymyd about hybrid quantum-classical computing, the new Maison du Quantique, and how real-world use cases may emerge over the next 5–7 years.