Tech
TAIWAN-TECH-BUSINESS-AI-COMPUTEX
(I-Hwa Cheng/Getty Images)
top of the flops

Hoppers, Blackwells, and Rubins: A field guide to the complicated world of Nvidia’s AI hardware

It’s common knowledge that Nvidia is at the core of the AI boom, but understanding what makes a “superchip” or why a NVL72 rack costs millions takes a bit of work.

Jon Keegan

No company has played a more central role to the current AI boom than Nvidia. It designed the chips, networking gear, and software that helped train today’s large language models and scale generative-AI products like ChatGPT to billions of users.

Understanding Nvidia’s AI hardware offerings, even for the tech savvy, can be challenging. While many of the biggest tech companies are hard at work building their own custom silicon to give them an edge in the ultracompetitive AI market, you will find Nvidia’s AI hardware powering pretty much every big AI data center out there today.

Some estimates have Nvidia owning as much as 98% of the data center GPU market. This has fueled the company’s meteoric rise to become one of the world’s largest companies. 

A chip by any other name...

To start understanding the landscape of Nvidia’s chips, it’s helpful to understand what each generation is called and which semis came out in that time. Going all the way back to 1999, Nvidia has named its various chip architectures after famous figures from science and mathematics. 

Earlier generations of Nvidia’s chip architecture powered the rise of advanced video graphics cards (in case you didn’t know, GPU stands for graphics processing unit) that helped propel the video game industry to new heights, but GPUs’ ability to run massively parallel vector math turned out to make them perfectly suited for AI.

The hot H100

The breakout star of Nvidia’s hardware offerings was undoubtedly the most powerful Hopper series chip, the H100 Tensor Core GPU. Announced in April 2022, this GPU was a breakthrough that featured the new “Transformer Engine,” a dedicated accelerator for the kinds of processing that large language models relied on for both training and “inference” (running a model) — which saw a 30x improvement from the previous generation’s fastest chip, the A100.

After OpenAI’s ChatGPT exploded onto the scene, demand for the H100 led tech companies to stockpile hoards of hundreds of thousands of the GPUs to help build bigger and faster large language models.

The H100s are estimated to cost between $20,000 and $40,000 each.

Nvidia H100
A Nvidia H100 GPU (Nvidia)

Blackwell “superchip”

In the fast-moving AI industry, while the H100 is still a hot item, the latest chip everyone is turning to is the GB200 — what Nvidia calls the “Grace Blackwell superchip.” This chip combines two Blackwell series B200 GPUs and a “Grace” CPU in one package.

Nvidia GB200 superchip
Nvidia CEO Jensen Huang holding a GB200 superchip at the Computex expo (Nvidia)

But if youre in the market for such powerful AI hardware, it’s likely you want dozens, hundreds, or even thousands of these chips wired up with the fastest interconnections you can get. That’s where the “GB200 NVL72” comes in. The NVL72 comes packed with 36 of the GB200 superchips — so 36 Grace CPUs and 72 of the B200 GPUs. Confused yet?

And if youre going on a GPU shopping spree, you better have lined up some VCs with deep pockets. Each GB200 superchip is estimated to cost between $60,000 and $70,000, while a fully equipped NVL72 rack is estimated to cost roughly $3 million, as it requires not only the pricey superchips but also expensive networking and liquid cooling.

If that’s too rich for you, you can always turn to AI investor darling CoreWeave, which advertises access to its batch of GB200 NVL72s starting at $42 per hour. CoreWeave says it has over 250,000 Nvidia GPUs in its data centers.

Chips within chips

According to Bloomberg, the “Stargate” mega data center project backed by OpenAI, SoftBank, and Oracle is planning on installing 400,000 of the GB200 superchips.

And Meta CEO Mark Zuckerberg has stated that he expects the company to have over 1.3 million GPUs by the end of 2025.

Leaps in performance

When youre talking about leaps forward in AI, its important to remember than rather than slow incremental bumps, each generation of chips is making exponential gains in a metric known as FLOPS, which measures performance.

Rubin matters

All this Nvidia jargon aside, there’s one model name you should pay attention to: Rubin, which will be the next leap forward in compute power.

Next year we’ll see the first of the Rubin architecture chips, the “Vera Rubin” superchip named after the American astronomer known for discovering dark matter.

Following the Vera Rubin chip release will be the Vera Rubin NVL144 (144 GPUs) and then Vera Rubin Ultra NVL576 (576 GPUs) in the second half of 2027.

Phew. Got all that?

More Tech

See all Tech
tech

Sora’s ghoulish reanimation of dead celebrities raises alarms

OpenAI’s video generation app Sora has spent its first two weeks at the top of the charts.

The startup’s fast-and-loose approach to enforcing intellectual property rights has seen the app flooded with videos of trademarked characters in all sorts of ugly scenarios.

But another area where Sora users have been pushing the limits involves videos that reanimate dead celebrities.

And we’re not talking just JFK, MLK, and Einstein. Videos featuring more recently deceased figures such as Robin Williams (11 years ago), painter Bob Ross (30 years ago), Stephen Hawking (seven years ago), and even Queen Elizabeth II (three years ago) have been generated. Some of the videos are racist and offensive, shocking the relatives of the figures.

OpenAI told The Washington Post that it is now allowing representatives of “recently deceased” celebrities and public figures to request that their likenesses be blocked from the service, though the company did not give a precise time frame for what it considered recent.

But another area where Sora users have been pushing the limits involves videos that reanimate dead celebrities.

And we’re not talking just JFK, MLK, and Einstein. Videos featuring more recently deceased figures such as Robin Williams (11 years ago), painter Bob Ross (30 years ago), Stephen Hawking (seven years ago), and even Queen Elizabeth II (three years ago) have been generated. Some of the videos are racist and offensive, shocking the relatives of the figures.

OpenAI told The Washington Post that it is now allowing representatives of “recently deceased” celebrities and public figures to request that their likenesses be blocked from the service, though the company did not give a precise time frame for what it considered recent.

tech

Tesla is selling unsold Cybertrucks to Elon Musk’s other companies

Sales of Tesla’s Cybertruck, once expected to reach hundreds of thousands per year, are currently in the low tens of thousands range and falling. Last quarter in the US, Tesla sold fewer than 5,400 of the “apocalypse-proof” vehicles, for a total of about 16,000 this year, Business Insider reports, citing Cox Automotive data.

That’s a 63% drop from the same quarter a year ago, even as Tesla as a whole notched its best quarterly sales ever, spurred by the expiration of the $7,500 federal EV tax credit.

With sales lagging, the company has dialed back production of the stainless steel behemoths, but there’s still been an excess.

Fortunately for Tesla, Electrek reports that CEO Elon Musk has other uses for Cybertrucks within his other companies, which often share resources and personnel. Tesla is delivering truckloads of the EV to both xAI (which Tesla shareholders will vote next month on whether to invest in) and SpaceX, where Cybertrucks are replacing internal combustion engine support fleets.

There’s a lot of chatter about “circular deals” in the billion-dollar pacts announced in the AI space on a weekly basis. But it doesn’t get much more circular than this, with production and buying activity kept within the Musk corporate family.

That’s a 63% drop from the same quarter a year ago, even as Tesla as a whole notched its best quarterly sales ever, spurred by the expiration of the $7,500 federal EV tax credit.

With sales lagging, the company has dialed back production of the stainless steel behemoths, but there’s still been an excess.

Fortunately for Tesla, Electrek reports that CEO Elon Musk has other uses for Cybertrucks within his other companies, which often share resources and personnel. Tesla is delivering truckloads of the EV to both xAI (which Tesla shareholders will vote next month on whether to invest in) and SpaceX, where Cybertrucks are replacing internal combustion engine support fleets.

There’s a lot of chatter about “circular deals” in the billion-dollar pacts announced in the AI space on a weekly basis. But it doesn’t get much more circular than this, with production and buying activity kept within the Musk corporate family.

tech

Tesla has begun selling the Model Y Standard in parts of Europe, where it has lots of cheaper competition

Days after rolling out “Standard” trim levels of its Model Y and Model 3 in the US, Tesla has started selling the Model Y Standard in some European countries. Standard Model Ys begin at about 40,000 euros, depending on the country, roughly 10,000 euros cheaper than the current Premium versions. In the US, Standard versions are about $5,000 cheaper than their souped-up peers. The model isn’t yet on sale in the UK or Ireland, where cars are driven on the left-hand side of the road.

While the Standard Teslas are cheaper, they pale in comparison to the many affordable EV options available in Europe, including those from China’s BYD, some of which start below 25,000 euros. CEO Elon Musk has called Europe the company’s “weakest market,” blaming the lack of approval for Tesla’s full self-driving technology for the shortfall.

Model 3 Standards don’t appear to be available yet in Europe.

tech
Jon Keegan

OpenAI commits up to $25 billion for 500-megawatt “Stargate Argentina” data center

OpenAI has reportedly signed a letter of intent to invest up to $25 billion on “Stargate Argentina,” a new 500-megawatt AI data center.

Reuters reports that the deal would involve tax incentives.

In a video announcing the project, OpenAI CEO Sam Altman said:

“Our vision for Stargate Argentina is to deliver a major boost to the country’s AI infrastructure, creating a foundation for new capabilities from smarter public services to tools that help small businesses compete globally.

OpenAI did not immediately respond to a request for comment.

You may remember the name “Stargate” from the megaproject that tech giants and the Trump administration announced earlier this year to build a huge number of data centers in the US. And you may remember Argentina as the nation the Trump administration is now bailing out with a $20 billion currency swap.

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, or Robinhood Money, LLC.