Tech
TAIWAN-TECH-BUSINESS-AI-COMPUTEX
(I-Hwa Cheng/Getty Images)
top of the flops

Hoppers, Blackwells, and Rubins: A field guide to the complicated world of Nvidia’s AI hardware

It’s common knowledge that Nvidia is at the core of the AI boom, but understanding what makes a “superchip” or why a NVL72 rack costs millions takes a bit of work.

Jon Keegan

No company has played a more central role to the current AI boom than Nvidia. It designed the chips, networking gear, and software that helped train today’s large language models and scale generative-AI products like ChatGPT to billions of users.

Understanding Nvidia’s AI hardware offerings, even for the tech savvy, can be challenging. While many of the biggest tech companies are hard at work building their own custom silicon to give them an edge in the ultracompetitive AI market, you will find Nvidia’s AI hardware powering pretty much every big AI data center out there today.

Some estimates have Nvidia owning as much as 98% of the data center GPU market. This has fueled the company’s meteoric rise to become one of the world’s largest companies. 

A chip by any other name...

To start understanding the landscape of Nvidia’s chips, it’s helpful to understand what each generation is called and which semis came out in that time. Going all the way back to 1999, Nvidia has named its various chip architectures after famous figures from science and mathematics. 

Earlier generations of Nvidia’s chip architecture powered the rise of advanced video graphics cards (in case you didn’t know, GPU stands for graphics processing unit) that helped propel the video game industry to new heights, but GPUs’ ability to run massively parallel vector math turned out to make them perfectly suited for AI.

The hot H100

The breakout star of Nvidia’s hardware offerings was undoubtedly the most powerful Hopper series chip, the H100 Tensor Core GPU. Announced in April 2022, this GPU was a breakthrough that featured the new “Transformer Engine,” a dedicated accelerator for the kinds of processing that large language models relied on for both training and “inference” (running a model) — which saw a 30x improvement from the previous generation’s fastest chip, the A100.

After OpenAI’s ChatGPT exploded onto the scene, demand for the H100 led tech companies to stockpile hoards of hundreds of thousands of the GPUs to help build bigger and faster large language models.

The H100s are estimated to cost between $20,000 and $40,000 each.

Nvidia H100
A Nvidia H100 GPU (Nvidia)

Blackwell “superchip”

In the fast-moving AI industry, while the H100 is still a hot item, the latest chip everyone is turning to is the GB200 — what Nvidia calls the “Grace Blackwell superchip.” This chip combines two Blackwell series B200 GPUs and a “Grace” CPU in one package.

Nvidia GB200 superchip
Nvidia CEO Jensen Huang holding a GB200 superchip at the Computex expo (Nvidia)

But if youre in the market for such powerful AI hardware, it’s likely you want dozens, hundreds, or even thousands of these chips wired up with the fastest interconnections you can get. That’s where the “GB200 NVL72” comes in. The NVL72 comes packed with 36 of the GB200 superchips — so 36 Grace CPUs and 72 of the B200 GPUs. Confused yet?

And if youre going on a GPU shopping spree, you better have lined up some VCs with deep pockets. Each GB200 superchip is estimated to cost between $60,000 and $70,000, while a fully equipped NVL72 rack is estimated to cost roughly $3 million, as it requires not only the pricey superchips but also expensive networking and liquid cooling.

If that’s too rich for you, you can always turn to AI investor darling CoreWeave, which advertises access to its batch of GB200 NVL72s starting at $42 per hour. CoreWeave says it has over 250,000 Nvidia GPUs in its data centers.

Chips within chips

According to Bloomberg, the “Stargate” mega data center project backed by OpenAI, SoftBank, and Oracle is planning on installing 400,000 of the GB200 superchips.

And Meta CEO Mark Zuckerberg has stated that he expects the company to have over 1.3 million GPUs by the end of 2025.

Leaps in performance

When youre talking about leaps forward in AI, its important to remember than rather than slow incremental bumps, each generation of chips is making exponential gains in a metric known as FLOPS, which measures performance.

Rubin matters

All this Nvidia jargon aside, there’s one model name you should pay attention to: Rubin, which will be the next leap forward in compute power.

Next year we’ll see the first of the Rubin architecture chips, the “Vera Rubin” superchip named after the American astronomer known for discovering dark matter.

Following the Vera Rubin chip release will be the Vera Rubin NVL144 (144 GPUs) and then Vera Rubin Ultra NVL576 (576 GPUs) in the second half of 2027.

Phew. Got all that?

More Tech

See all Tech
tech
Rani Molla

Meta reportedly strikes multibillion-dollar AI chip deal with Google as it struggles to design its own

Meta has signed a deal with Google to rent tensor processing units to develop new AI models and is in talks to buy the chips for its data centers, The Information reports.

The agreement comes on top of a recently announced “multi-generational” partnership with Nvidia and a chip supply deal with Advanced Micro Devices that could be worth more than $100 billion, as Meta scrapped its most advanced in-house AI training chip amid design challenges.

A Meta deal with Google, which has been rumored since November, would position the search giant more directly as a competitor to Nvidia in its core business of AI processors. Some analysts have said selling its custom chips to outside customers could become a business worth hundreds of billions of dollars for Google.

A Meta deal with Google, which has been rumored since November, would position the search giant more directly as a competitor to Nvidia in its core business of AI processors. Some analysts have said selling its custom chips to outside customers could become a business worth hundreds of billions of dollars for Google.

tech
Jon Keegan

Delays in permitting, power, and zoning cause first drop in data center construction since 2020

Despite incredible demand, the number of data centers under construction in North America fell for the first time since 2020, according to new research from CBRE.

Total data center capacity under construction dropped about 5.6% year on year from 6.35 megawatts in 2024 to 5.99 megawatts by the end of 2025.

What’s causing the delay? Slow permitting, constrained supply chains, and growing public engagement with how deals are approved at the local level. Labor constraints also were cited in the report; a tight supply of skilled workers will increase costs.

What’s causing the delay? Slow permitting, constrained supply chains, and growing public engagement with how deals are approved at the local level. Labor constraints also were cited in the report; a tight supply of skilled workers will increase costs.

-13%📱
Rani Molla

Smartphone shipments are expected to decline 13% — the biggest drop ever — to 1.12 billion in 2026, according to new data from IDC, as the memory shortage drives up costs and prices for phones. The firm expects the average smartphone selling price to jump 14% to a record $523 this year.

The shortfall will mostly affect makers of lower-end smartphones, whose customers are more cost-conscious, while higher-end manufacturers like Samsung and Apple are likely to be more insulated from the pressure.

“The memory crisis will cause more than a temporary decline; it marks a structural reset of the entire market, fundamentally reshaping long‑term TAM (Total Addressable Market), the vendor landscape, and the product mix,” said Nabila Popal, senior research director with IDCs Worldwide Quarterly Mobile Phone Tracker. “We expect consolidation as smaller players exit, and low-end vendors to face sharp shipment declines amid supply constraints and lower demand at higher price points.”

tech
Jon Keegan

Google drops new Nano Banana

Google is hoping to recapture the viral boost it received when it released its Nano Banana image generation model. Nano Banana 2 arrives today, which Google has rolled into its Gemini app.

The new model promises more accurate text rendering and translation and “advanced world knowledge,” which “pulls from Gemini’s real-world knowledge base, and is powered by real-time information and images from web search to more accurately render specific subjects,” according to the company’s press release.

New creative controls let users keep groups of characters consistent across scenes, render images with higher resolution, and parse complex prompts.

The first version of Nano Banana became popular for making action figures out of users, and helped catapult the Gemini AI app to the top of the charts, bumping ChatGPT from its perch.

New creative controls let users keep groups of characters consistent across scenes, render images with higher resolution, and parse complex prompts.

The first version of Nano Banana became popular for making action figures out of users, and helped catapult the Gemini AI app to the top of the charts, bumping ChatGPT from its perch.

tech
Rani Molla

Tesla’s ride-hailing service is looking a lot more like Uber’s than Waymo’s

Despite numerous promises about amassing a giant network of driverless cars, so far it seems like Tesla’s Robotaxis are a lot more similar to Uber’s plain old ride-hailing service than Waymo’s expanding autonomous fleet.

In California, where Tesla has its largest ride-hailing service, the company has taken no formal steps to gain approval for a truly driverless car service, according to Reuters. Throughout 2025, Tesla failed to log a single mile of autonomous test driving on state roads, and has not applied for the necessary permits to test or deploy vehicles without a human present. Currently, Tesla holds only a basic permit that requires a human safety monitor to remain in the driver’s seat at all times.

Currently, Tesla’s California Robotaxi service consists of roughly 300 Teslas operated by human drivers using the company’s supervised Full Self-Driving tech. In Austin, where the company has about 45 vehicles, Tesla made a big show earlier this year of announcing it was removing the safety monitors sitting in the front seats during rides. However, to date, only a handful of those vehicles have been reported to be actually operating without a safety monitor onboard.

In other words, it’s performing a service more akin to a tech-heavy Uber ride than the one operated by Alphabet subsidiary Waymo, which earlier this week announced it now has driverless rides available to the public in 10 markets. Even Uber is trying to put space between itself and the old driver-having Ubers of yore: this week its autonomous software partner said the company plans to launch a driverless service in London this year, with plans for 10 markets.

During its earnings report last month, Tesla said it planned to offer Robotaxi service in a half dozen new cities in the first half of this year, including Phoenix, Miami, and Las Vegas. Judging by Tesla’s progress so far, it’s likely those services will also feature a human in the front seat.

In California, where Tesla has its largest ride-hailing service, the company has taken no formal steps to gain approval for a truly driverless car service, according to Reuters. Throughout 2025, Tesla failed to log a single mile of autonomous test driving on state roads, and has not applied for the necessary permits to test or deploy vehicles without a human present. Currently, Tesla holds only a basic permit that requires a human safety monitor to remain in the driver’s seat at all times.

Currently, Tesla’s California Robotaxi service consists of roughly 300 Teslas operated by human drivers using the company’s supervised Full Self-Driving tech. In Austin, where the company has about 45 vehicles, Tesla made a big show earlier this year of announcing it was removing the safety monitors sitting in the front seats during rides. However, to date, only a handful of those vehicles have been reported to be actually operating without a safety monitor onboard.

In other words, it’s performing a service more akin to a tech-heavy Uber ride than the one operated by Alphabet subsidiary Waymo, which earlier this week announced it now has driverless rides available to the public in 10 markets. Even Uber is trying to put space between itself and the old driver-having Ubers of yore: this week its autonomous software partner said the company plans to launch a driverless service in London this year, with plans for 10 markets.

During its earnings report last month, Tesla said it planned to offer Robotaxi service in a half dozen new cities in the first half of this year, including Phoenix, Miami, and Las Vegas. Judging by Tesla’s progress so far, it’s likely those services will also feature a human in the front seat.

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, or Robinhood Money, LLC.