Microarchitecture Blackwell

Blackwell is a graphics processing unit (GPU) microarchitecture developed by Nvidia as the successor to the Hopper and Ada Lovelace microarchitectures.

Blackwell
LaunchedMarch 18, 2024 (2024-03-18)
Designed byNvidia
Manufactured by
Fabrication processTSMC 4NP
Codename(s)GB100
GB20x
Specifications
Memory supportHBM3e
PCIe supportPCIe 6.0
Supported Graphics APIs
DirectXDirectX 12 Ultimate (Feature Level 12_2)
Direct3DDirect3D 12
Shader ModelShader Model 6.8
OpenCLOpenCL 3.0
OpenGLOpenGL 4.6
CUDACompute Capability 10.x
VulkanVulkan 1.3
Supported Compute APIs
CUDACUDA Toolkit 10.0
DirectComputeYes
Media Engine
Encoder(s) supportedNVENC
History
PredecessorAda Lovelace (consumer)
Hopper (datacenter)

Named after statistician and mathematician David Blackwell, the name of the Blackwell architecture was leaked in 2022 with the B40 and B100 accelerators being confirmed in October 2023 with an official Nvidia roadmap shown during an investors presentation. and was officially announced at Nvidia's GTC 2024 keynote on March 18, 2024.

History

Microarchitecture Blackwell 
David Blackwell (1919-2010), eponym of architecture

In March 2022, Nvidia announced Hopper architecture for datacenter for AI accelerators. Demand for Hopper products was high throughout 2023's AI hype. The lead time from order to delivery of H100-based servers was between 36 and 52 weeks due to shortages and high demand. Nvidia reportedly sold 500,000 Hopper-based H100 accelerators in Q3 2023 alone. Nvidia's AI dominance with Hopper products led to the company increasing its market capitalization to over $2 trillion, behind only Microsoft and Apple.

The Blackwell architecture is named after American mathematician David Blackwell who was known for his contributions to the mathematical fields of game theory, probability theory, information theory, and statistics. These areas have influenced or are implemented in transformer-based generative AI model designs or their training algorithms. Blackwell was the first African American scholar to be inducted into the National Academy of Sciences.

In Nvidia's October 2023 Investor Presentation, its datacenter roadmap was updated to include reference to its B100 and B40 accelerators and the Blackwell architecture. Previously, the successor to Hopper was simply named on roadmaps as "Hopper-Next". Nvidia's updated roadmap emphasized the move from a two-year release cadence for datacenter products to yearly releases targeted for x86 and ARM systems.

At the Graphics Technology Conference (GTC) on March 18, 2024, Nvidia officially announced the Blackwell architecture with focus placed on its B100 and B200 datacenter accelerators. Nvidia CEO Jensen Huang said that with Blackwell, "we created a processor for the generative AI era" and emphasized the overall Blackwell platform combining Blackwell accelerators with Nvidia's ARM-based Grace CPU. Nvidia touted endorsements of Blackwell from the CEOs of Google, Meta, Microsoft, OpenAI and Oracle. The keynote did not mention gaming.

Architecture

Blackwell is an architecture designed for both datacenter compute applications and for gaming and workstation applications with dedicated dies for each purpose. The GB100 die is for Blackwell datacenter products while GB200 series dies will be used for GeForce RTX 50 series graphics cards.

Process node

Blackwell is fabricated on the custom 4NP node from TSMC. 4NP is an enhancement of the 4N node used for the Hopper and Ada Lovelace architectures with an increase in transistor density. With the enhanced 4NP node, the GB100 die contains 104 billion transistors, a 30% increase over the 80 billion transistors in the previous generation Hopper GH100 die. As Blackwell cannot reap the benefits that come with a major process node advancement, it must achieve power efficiency and performance gains through underlying architectural changes.

The GB100 die is at the reticle limit of semiconductor fabrication. The reticle limit in semiconductor fabrication is the physical size limit that lithography machines can etch a silicon die. Previously, Nvidia had nearly hit TSMC's reticle limit with GH100's 814mm2 die. In order to not be constrained by die size, Nvidia's B100 accelerator utilizes two GB100 dies in a single package, connected with a 10 TB/s link that Nvidia calls the NV-High Bandwidth Interface (NV-HBI). NV-HBI is based on the NVLink 5.0 protocol. Nvidia CEO Jensen Huang claimed in an interview with CNBC that Nvidia had spent around $10 billion in research and development for Blackwell's NV-HBI die interconnect. Veteran semiconductor engineer Jim Keller, who had worked on AMD's K7, K12 and Zen architectures, criticized this figure and claimed that the same outcome could be achieved for $1 billion through using Ultra Ethernet rather than the proprietry NVLink system. The two connected GB100 dies are able to act like a large monolithic piece of silicon with full cache coherency between both dies. The dual die package totals 208 billion transistors. Those two GB100 dies are placed on top on a silicon interposer produced using TSMC's CoWoS-L 2.5D packaging technique.

Streaming Multiprocessor

CUDA Cores

CUDA Compute Capability 10.0 is added with Blackwell.

Tensor Cores

The Blackwell architecture introduces fifth generation Tensor Cores for AI compute and performing floating-point calculations. In the datacenter, Blackwell adds support for FP4 and FP6 data types with eighth precision floating point processing. The previous Hopper architecture introduced the Transformer Engine to divide FP32 data into FP8 to increase peak compute throughput. Blackwell's second generation Transformer Engine allows FP32 to be divided further, allow a doubling of FP8 compute performance. Using 4-bit data allows greater efficiency and throughput for model inference during generative AI training. Nvidia claims 20 petaflops of FP4 compute with the dual GB100 die B100 accelerator.

See also

References

Tags:

Microarchitecture Blackwell HistoryMicroarchitecture Blackwell ArchitectureMicroarchitecture Blackwell

🔥 Trending searches on Wiki English:

The Three-Body Problem (novel)The Beekeeper (2024 film)Skibidi ToiletTraumatic inseminationJay-ZCamila CabelloItalyAmazon (company)U.S. stateJohn EastmanList of bridge failuresMckenna GraceSandro TonaliEva MendesSuge KnightDead Internet theoryPirates of the Caribbean (film series)Tyla (South African singer)Solo LevelingEngland national football teamAmerican Civil WarOlivia RodrigoSatyadeep MishraCoral CastleRonna McDanielMumbai IndiansJohnny CashDark forest hypothesisFrancesca AlbaneseJennifer AnistonThree-body problemRobloxRuth KearneyRaindrop cakeList of countries by GDP (nominal)Charlie SheenFermi paradoxEid al-FitrKatt WilliamsElvis PresleyHoliKim Ji-won (actress)BuddhismJoe LiebermanTheodore RooseveltRicky StanickyYellowstone (American TV series)Jon Paul SteuerMel BJulius CaesarNatalie PortmanAndre Jin CoquillardHarold RamisJanet JacksonTiger WoodsHenry CavillMaura TierneyGeorge VNirmala SitharamanChennai Super KingsLast SupperPost MaloneKalanithi MaranThe Gentlemen (2019 film)The Accountant (2016 film)New York CityThe Ministry of Ungentlemanly WarfareWes MooreMichael SchumacherSawai Mansingh StadiumJulian AssangeRobert HanssenRosalind ChaoShirley ChisholmNarendra ModiFrançois TruffautTokyo Vice (TV series)🡆 More