
At the Consumer Electronics Show, Nvidia's CEO Jensen Huang announced the launch of the company's latest Rubin computing architecture, heralded as a breakthrough in artificial intelligence hardware. This advanced architecture is currently in production and is expected to ramp up significantly later this year. Huang emphasized the critical need for more computational power, stating, "The amount of computation necessary for AI is skyrocketing," and assured attendees that Rubin is now fully in production. Initially revealed in 2024, the Rubin architecture signifies Nvidia’s ongoing commitment to hardware innovation, reinforcing its position as the world's most valuable corporation. Replacing the previous Blackwell architecture, which itself succeeded Hopper and Lovelace, the Rubin architecture is set to be utilized by prominent cloud service providers. Major partnerships with industry leaders such as Anthropic, OpenAI, and Amazon Web Services will see Rubin chips powering their infrastructures. Additionally, the architecture will contribute to the HPE Blue Lion supercomputer and the upcoming Doudna supercomputer at Lawrence Berkeley National Lab. Named after the renowned astronomer Vera Florence Cooper Rubin, this architecture comprises six integrated chips working together seamlessly. The centerpiece is the Rubin GPU, complemented by enhancements in storage and interconnection via the new Bluefield and NVLink systems. A notable aspect of the architecture is the introduction of the Vera CPU, tailored for agentic reasoning tasks. Dion Harris, Nvidia’s senior director of AI infrastructure solutions, highlighted the innovative storage capabilities designed to meet the increasing memory demands of contemporary AI workflows. He mentioned, "As you start to enable new types of workflows, like agentic AI or long-term tasks, that puts a lot of stress and requirements on your KV cache." The new tier of storage allows for more efficient scaling of storage pools, connecting externally to the compute device. The Rubin architecture also boasts remarkable advancements in speed and energy efficiency. Nvidia’s evaluations indicate that it will outperform the Blackwell architecture by three and a half times in model-training tasks and achieve five times the speed in inference tasks, reaching an impressive 50 petaflops. Furthermore, it will provide eight times more inference compute per watt. Amid fierce competition in the AI infrastructure sector, Huang projected that spending on AI infrastructure could reach between $3 trillion and $4 trillion over the next five years.
Fi, the Indian neobank that gained traction over the past four years, is officially winding down its banking operations....
TechCrunch | Mar 11, 2026, 22:30
Nuro, a startup from Silicon Valley backed by prominent investors including Nvidia, Uber, and Softbank, is stepping into...
TechCrunch | Mar 11, 2026, 23:35
In a significant development last week, Netflix revealed its acquisition of InterPositive, an innovative AI company co-f...
TechCrunch | Mar 11, 2026, 22:30
In a pointed critique, Federal Communications Commission Chairman Brendan Carr called out Amazon for its opposition to S...
CNBC | Mar 11, 2026, 22:30
The recent surge in artificial intelligence spending is transforming the memory industry in unprecedented ways. Over the...
CNBC | Mar 11, 2026, 21:15