The Next Platform
  • Home
  • Compute
  • Store
  • Connect
  • Control
  • Code
  • AI
  • HPC
  • Enterprise
  • Hyperscale
  • Cloud
  • Edge
Latest
  • [ October 21, 2025 ] Software Pushes The AI Pareto Frontier More Than Hardware AI
  • [ October 20, 2025 ] “Polaris” AmpereOne M Arm CPUs Sighted In Oracle A4 Instances Compute
  • [ October 17, 2025 ] The Third Time Will Be The Charm For Broadcom Switch Co-Packaged Optics Connect
  • [ October 17, 2025 ] How HPC Is Igniting Discoveries In Dinosaur Locomotion – And Beyond HPC
  • [ October 16, 2025 ] TSMC Is Running Ahead Of Forecasts On AI Growth Compute
  • [ October 15, 2025 ] Ellison: Oracle Is Leveraging All Of Its Advantages To Build A Different Cloud AI
  • [ October 14, 2025 ] Oracle First In Line For AMD “Altair” MI450 GPUs, “Helios” Racks Compute
  • [ October 14, 2025 ] Google Rolls Up Gemini And AI Tools Into An Enterprise Platform AI
HomeRubin CPX

Rubin CPX

Compute

Nvidia Disaggregates Long-Context Inference To Drive Bang For The Buck

September 11, 2025 Timothy Prickett Morgan 5

It is beginning to look like that the period spanning from the second half of 2026 through the first half of 2027 is going to be a local maximum in spending on XPU-accelerated systems for AI workloads. …

About

The Next Platform is part of the Situation Publishing family, which includes the enterprise and business technology publication, The Register.

TNP  offers in-depth coverage of high-end computing at large enterprises, supercomputing centers, hyperscale data centers, and public clouds. Read more…

Newsletter

Featuring highlights, analysis, and stories from the week directly from us to your inbox with nothing in between.
Subscribe now

  • RSS
  • Twitter
  • Facebook
  • LinkedIn
  • Email the editor
  • About
  • Contributors
  • Contact
  • Sales
  • Newsletter
  • Books
  • Events
  • Privacy
  • Ts&Cs
  • Cookies
  • Do not sell my personal information

All Content Copyright The Next Platform