
Jensen Huang reveals the Blackwell architecture, promising massive efficiency gains for LLM training and inference.
The hardware landscape of artificial intelligence is undergoing a seismic shift with the official production ramp-up of NVIDIA's Blackwell architecture. Named after mathematician David Blackwell, the new platform is designed to handle the trillion-parameter scale of future large language models. The flagship GB200 Grace Blackwell Superchip connects two Blackwell GPUs to a Grace CPU, offering a 30x performance increase for LLM inference workloads compared to the previous H100 generation.
Beyond raw speed, NVIDIA is emphasizing the energy efficiency of the new chips. As data centers face increasing pressure regarding their carbon footprint, Blackwell claims to reduce energy consumption and costs by up to 25 times. This is achieved through a dedicated engine that dynamically manages precision levels, allowing the hardware to utilize lower-bit arithmetic where high precision is not required for the final output.
Major cloud providers including AWS, Microsoft Azure, and Google Cloud have already announced plans to integrate Blackwell-based clusters into their infrastructure. As the industry moves toward agentic AI systems that require constant background computation, the hardware efficiency provided by Blackwell will likely become the standard for scaling global AI services.


