Tech
TAIWAN-TECH-BUSINESS-AI-COMPUTEX
(I-Hwa Cheng/Getty Images)
top of the flops

Hoppers, Blackwells, and Rubins: A field guide to the complicated world of Nvidia’s AI hardware

It’s common knowledge that Nvidia is at the core of the AI boom, but understanding what makes a “superchip” or why a NVL72 rack costs millions takes a bit of work.

Jon Keegan

No company has played a more central role to the current AI boom than Nvidia. It designed the chips, networking gear, and software that helped train today’s large language models and scale generative-AI products like ChatGPT to billions of users.

Understanding Nvidia’s AI hardware offerings, even for the tech savvy, can be challenging. While many of the biggest tech companies are hard at work building their own custom silicon to give them an edge in the ultracompetitive AI market, you will find Nvidia’s AI hardware powering pretty much every big AI data center out there today.

Some estimates have Nvidia owning as much as 98% of the data center GPU market. This has fueled the company’s meteoric rise to become one of the world’s largest companies. 

A chip by any other name...

To start understanding the landscape of Nvidia’s chips, it’s helpful to understand what each generation is called and which semis came out in that time. Going all the way back to 1999, Nvidia has named its various chip architectures after famous figures from science and mathematics. 

Earlier generations of Nvidia’s chip architecture powered the rise of advanced video graphics cards (in case you didn’t know, GPU stands for graphics processing unit) that helped propel the video game industry to new heights, but GPUs’ ability to run massively parallel vector math turned out to make them perfectly suited for AI.

The hot H100

The breakout star of Nvidia’s hardware offerings was undoubtedly the most powerful Hopper series chip, the H100 Tensor Core GPU. Announced in April 2022, this GPU was a breakthrough that featured the new “Transformer Engine,” a dedicated accelerator for the kinds of processing that large language models relied on for both training and “inference” (running a model) — which saw a 30x improvement from the previous generation’s fastest chip, the A100.

After OpenAI’s ChatGPT exploded onto the scene, demand for the H100 led tech companies to stockpile hoards of hundreds of thousands of the GPUs to help build bigger and faster large language models.

The H100s are estimated to cost between $20,000 and $40,000 each.

Nvidia H100
A Nvidia H100 GPU (Nvidia)

Blackwell “superchip”

In the fast-moving AI industry, while the H100 is still a hot item, the latest chip everyone is turning to is the GB200 — what Nvidia calls the “Grace Blackwell superchip.” This chip combines two Blackwell series B200 GPUs and a “Grace” CPU in one package.

Nvidia GB200 superchip
Nvidia CEO Jensen Huang holding a GB200 superchip at the Computex expo (Nvidia)

But if youre in the market for such powerful AI hardware, it’s likely you want dozens, hundreds, or even thousands of these chips wired up with the fastest interconnections you can get. That’s where the “GB200 NVL72” comes in. The NVL72 comes packed with 36 of the GB200 superchips — so 36 Grace CPUs and 72 of the B200 GPUs. Confused yet?

And if youre going on a GPU shopping spree, you better have lined up some VCs with deep pockets. Each GB200 superchip is estimated to cost between $60,000 and $70,000, while a fully equipped NVL72 rack is estimated to cost roughly $3 million, as it requires not only the pricey superchips but also expensive networking and liquid cooling.

If that’s too rich for you, you can always turn to AI investor darling CoreWeave, which advertises access to its batch of GB200 NVL72s starting at $42 per hour. CoreWeave says it has over 250,000 Nvidia GPUs in its data centers.

Chips within chips

According to Bloomberg, the “Stargate” mega data center project backed by OpenAI, SoftBank, and Oracle is planning on installing 400,000 of the GB200 superchips.

And Meta CEO Mark Zuckerberg has stated that he expects the company to have over 1.3 million GPUs by the end of 2025.

Leaps in performance

When youre talking about leaps forward in AI, its important to remember than rather than slow incremental bumps, each generation of chips is making exponential gains in a metric known as FLOPS, which measures performance.

Rubin matters

All this Nvidia jargon aside, there’s one model name you should pay attention to: Rubin, which will be the next leap forward in compute power.

Next year we’ll see the first of the Rubin architecture chips, the “Vera Rubin” superchip named after the American astronomer known for discovering dark matter.

Following the Vera Rubin chip release will be the Vera Rubin NVL144 (144 GPUs) and then Vera Rubin Ultra NVL576 (576 GPUs) in the second half of 2027.

Phew. Got all that?

More Tech

See all Tech
tech

California AG launches probe into xAI and Grok over sexualized deepfakes of women and children

The California attorney general just opened an investigation into xAI, Elon Musk’s AI startup, over chatbot Grok’s apparent role in generating nonconsensual sexual images of women and children. The probe centers on reports that Grok has been used to facilitate the creation of sexually explicit images without consent, many of which have circulated on X.

“The avalanche of reports detailing the non-consensual, sexually explicit material that xAI has produced and posted online in recent weeks is shocking,” Attorney General Rob Bonta wrote in a press release. “As the top law enforcement official of California tasked with protecting our residents, I am deeply concerned with this development in AI and will use all the tools at my disposal to keep California’s residents safe.”

California’s move follows growing scrutiny from U.S. lawmakers and the UK government over AI-generated sexual content and deepfakes.

xAI and Tesla CEO Musk earlier today wrote, “I not aware of any naked underage images generated by Grok. Literally zero.”

Grok is currently No. 5 on Apple’s free App Store.

“The avalanche of reports detailing the non-consensual, sexually explicit material that xAI has produced and posted online in recent weeks is shocking,” Attorney General Rob Bonta wrote in a press release. “As the top law enforcement official of California tasked with protecting our residents, I am deeply concerned with this development in AI and will use all the tools at my disposal to keep California’s residents safe.”

California’s move follows growing scrutiny from U.S. lawmakers and the UK government over AI-generated sexual content and deepfakes.

xAI and Tesla CEO Musk earlier today wrote, “I not aware of any naked underage images generated by Grok. Literally zero.”

Grok is currently No. 5 on Apple’s free App Store.

tech

Report: Microsoft on track to spend $500 million per year on Anthropic AI

Last fall, Microsoft and OpenAI’s $13 billion partnership seemed to finally be on solid ground.

OpenAI’s restructuring was completed on time, and the companies hammered out an updated agreement that secured OpenAI’s status as Microsoft’s AI provider of choice, but also allowed for Microsoft to work with other companies.

Now Microsoft is doing exactly that. Microsoft has been increasing its spending on Anthropic’s AI, and is on track to spend $500 million per year on the startup’s services, according to a new report from The Information.

The increasingly cozy relationship between the companies includes the rare move of Microsoft offering incentives to its salespeople that allows Anthropic sales to count toward their quotas, per to the report. Microsoft invested $5 billion in Anthropic as part of a big deal in November that included Nvidia.

Microsoft has also been using Anthropic’s AI to power more and more of its own products, such as Github Copilot and 365 Copilot.

Now Microsoft is doing exactly that. Microsoft has been increasing its spending on Anthropic’s AI, and is on track to spend $500 million per year on the startup’s services, according to a new report from The Information.

The increasingly cozy relationship between the companies includes the rare move of Microsoft offering incentives to its salespeople that allows Anthropic sales to count toward their quotas, per to the report. Microsoft invested $5 billion in Anthropic as part of a big deal in November that included Nvidia.

Microsoft has also been using Anthropic’s AI to power more and more of its own products, such as Github Copilot and 365 Copilot.

tech

Report: Apple staggers Siri AI rollout, with key features pushed to summer

Thanks to Apple’s new partnership with Google, the Gemini-backed version of Siri should begin rolling out this spring, but several key features Apple previewed in 2024 may not come until summer, The Information reports.

The new Siri is soon expected to answer general questions with ChatGPT-like answers — rather than quoting directly from websites or not answering at all. But more personalized, proactive features, like, for example, remembering past conversations and information from them to suggest you leave for a planned trip earlier to beat traffic, may not be unveiled until June at the company’s Worldwide Developers Conference.

The report also clarifies that while Apple’s partnership with Microsoft-backed OpenAI, wherein users could summon ChatGPT for complex questions, isn’t changing, the Google deal might reduce the need for people to do so because Siri will likely be able to answer those questions itself. The Information notes, citing a person familiar with the deal, that the ChatGPT option hadn’t driven much traffic to OpenAI before.

The new Siri is soon expected to answer general questions with ChatGPT-like answers — rather than quoting directly from websites or not answering at all. But more personalized, proactive features, like, for example, remembering past conversations and information from them to suggest you leave for a planned trip earlier to beat traffic, may not be unveiled until June at the company’s Worldwide Developers Conference.

The report also clarifies that while Apple’s partnership with Microsoft-backed OpenAI, wherein users could summon ChatGPT for complex questions, isn’t changing, the Google deal might reduce the need for people to do so because Siri will likely be able to answer those questions itself. The Information notes, citing a person familiar with the deal, that the ChatGPT option hadn’t driven much traffic to OpenAI before.

Mark Zuckerberg in the metaverse

RIP the metaverse

Meta seems to be winding down its metaverse ambitions. We took a look back at what the company was going for.

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, or Robinhood Money, LLC.