Tech
Full frame Close up Background Blueberries
(Sunlight7/Getty Images)

GPT-5: “A legitimate PhD level expert in anything” that sucks at spelling and geography

OpenAI spent a lot of time talking about how “smart” GPT-5 is, yet it failed spectacularly at tasks that a second grader could achieve.

Jon Keegan

Yesterday, OpenAI spent a lot of time talking about how “smart” its new GPT-5 model was.

OpenAI cofounder and CEO Sam Altman said of GPT-5: “Its like talking to an expert, a legitimate PhD level expert in anything,” and called it “the most powerful, the most smart, the fastest, the most reliable and the most robust reasoning model that weve shipped to date.”

Demos showed GPT-5 effortlessly creating an interactive simulation to explain the Bernoulli effect, diagnosing and fixing complex code errors, planning personalized running schedules, and “vibecoding” a cartoony 3D castle game. The company touted benchmarks showing how GPT-5 aced questions from a Harvard-MIT mathematics tournament and got high scores on coding, visual problem-solving, and other impressive tasks.

But once the public got its chance to kick GPT-5’s tires, some cracks started to emerge in this image of a superintelligent expert in everything.

AI models are famously bad at spelling different kinds of berries, and GPT-5 is no exception.

I had to try the “blueberry” thing myself with GPT5. I merely report the results.

[image or embed]

— Kieran Healy (@kjhealy.co) August 7, 2025 at 8:04 PM

Another classic failure of being bad at maps persisted with GPT-5 when it was asked to list all the states with the letter R in their names. It even offered to map them, with hilarious results:

My goto is to ask LLMs how many states have R in their name. They always fail. GPT 5 included Indiana, Illinois, and Texas in its list. It then asked me if I wanted an alphabetical highlighted map. Sure, why not.

[image or embed]

— radams (@radamssmash.bsky.social) August 7, 2025 at 8:40 PM

To be fair to OpenAI, this problem isn’t unique to GPT-5, as similar failures were documented with Google’s Gemini.

During the livestream, in an embarrassing screwup, a presentation showed some charts that looked like they had some of the same problems.

If this was all described as a “technical preview,” these kinds of mistakes might be understandable. But this is a real product from a company thats pulling in $1 billion per month. OpenAI’s models are being sold to schools, hospitals, law firms, and government agencies, including for national security and defense.

OpenAI is also telling users that GPT-5 can be used for medical information, while cautioning that the model does not replace a medical professional.

“GPT‑5 is our best model yet for health-related questions, empowering users to be informed about and advocate for their health.”

Why is this so hard?

The reason why such an advanced model can appear to be so capable at complex coding, math, and physics yet fail so spectacularly at spelling and maps is that generative models like GPT-5 are probabilistic systems at their core — they predict the most likely next token based on the volumes of data they have been trained on. They don’t know anything, and they don’t think or have any model of how the world works (though researchers are working on that).

When the model writes its response, you see the thing that has the highest score for what you should see next. But with math and coding, the rules are more strict and the examples its been trained on are consistent, so it has higher accuracy and can ace the math benchmarks with flying colors.

But drawing a map with names or counting the letters in a word are weirdly tough, as it requires a skill the model doesn’t really have and has to figure out step by step from patterns, which can lead to odd results. That’s a simplification of a very complex and sophisticated system, but applies to a lot of the generative-AI technology in use today.

Thats also a whole lot to explain to users, but OpenAI boils those complicated ideas down to a single warning below the text box: “ChatGPT can make mistakes. Check important info.”

OpenAI did not immediately respond to a request for comment.

More Tech

See all Tech
tech
Jon Keegan

Judge blocks Pentagon’s move to blacklist Anthropic

A federal judge in Northern California has granted a preliminary injunction blocking the Pentagon from labeling Anthropic as a national security supply chain risk.

The ruling temporarily prevents the Defense Department from restricting the AI company’s access to federal contracts amid a dispute over its refusal to allow certain military and surveillance uses of its technology. The designation could also have shifted lucrative government work toward competitors, including OpenAI.

Earlier this month, Anthropic, the company behind Claude, sued 17 federal agencies and their heads, alleging the government exceeded its statutory authority.

tech
Rani Molla

Report: SpaceX’s record IPO may grant preferential access to retail investors and Tesla shareholders

SpaceX’s impending IPO could raise $40 billion to $80 billion and rank as the largest ever — as well as one of the most unconventional.

The Wall Street Journal reports several ways CEO Elon Musk is considering breaking with IPO norms:

  • Investors in his other companies, including Tesla, could receive preferential access to shares.

  • Individual investors may get a third or more of the allocation, far above the typical ~10% mark.

  • Instead of a traditional road show, Musk wants investors to visit SpaceX facilities in person.

  • Investors in his other companies, including Tesla, could receive preferential access to shares.

  • Individual investors may get a third or more of the allocation, far above the typical ~10% mark.

  • Instead of a traditional road show, Musk wants investors to visit SpaceX facilities in person.

tech
Rani Molla

Tesla released estimates for Q1 deliveries and they’re lower than analysts expected

Ahead of first-quarter earnings next month, Tesla released its own company-compiled Wall Street consensus estimate for deliveries: 365,645 vehicles. While that’s lower than the 382,000 FactSet consensus estimate, it represents a nearly 9% jump from Q1 2025, when Tesla sold 336,681 vehicles.

Tesla started releasing its own consensus estimates to the public — not just institutional investors — for the first time in Q4 2025. The move was seen as a way to temper investor expectations, as other estimates were too high. Last quarter, Tesla’s compilation was closer to actual numbers, which fell 16% year over year.

The market-implied odds from event contracts suggest 64% of traders think Tesla’s Q1 deliveries will be more than 350,000, 44% think it will be higher than 360,000, and just 21% have it at higher than 370,000.

(Event contracts are offered through Robinhood Derivatives, LLC — probabilities referenced or sourced from KalshiEx LLC or ForecastEx LLC.)

ARC-AGI-3

The toughest AI benchmark just got a whole lot tougher

ARC-AGI-3 is the latest version of a clever benchmark that challenges AI models to solve mini video games with no written instructions.

Jon Keegan3/26/26

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, Robinhood Derivatives, LLC, or Robinhood Money, LLC. Futures and event contracts are offered through Robinhood Derivatives, LLC.