pull down to refresh

The company unveiled Rubin CPX, a graphics processor optimized for tasks with a context of over 1 million tokens.

The chip is designed for “disaggregated inference,” an approach where different GPUs process different parts of a task. This should improve the efficiency of models in video generation, programming, and other long-context scenarios.

Rubin CPX will be released in late 2026.

I read this sentence 5 times:

To address this shift, the NVIDIA SMART framework provides a path forward—optimizing inference across scale, multidimensional performance, architecture, ROI, and the broader technology ecosystem.

And I then decided that you don't need AI for slop. lol.

reply