Tech
Open AI Chief Executive Officer Sam Altman
(Kim Jae-Hwan/Getty Images)

OpenAI’s o1-pro is the most expensive AI model in the industry

No other model by a major AI company comes even close.

OpenAI just released the pricing for its o1-pro reasoning model, which draws on an insane amount of computing power to use multistep “reasoning” through problems to get better responses to prompts. This computing power doesn’t come cheap: the new pricing is the highest for any major model in the industry today, and by a lot.

As a regular human user, you can use a lot of AI tools for free, but maybe you pay $20 per month for OpenAI’s ChatGPT Plus or Google’s Gemini Advanced if you use it a lot. But that’s not where the money is.

When companies are hooking their services up to AI platforms behind the scenes via an API (application programming interface), the costs can really add up. So what is the standard unit of measure for AI costs?

API pricing for AI models is measured by how much data (words, images, video, audio) you put into a model and how much data gets spit back out to you. The output costs more than the input.

The common measure for this is 1 million “tokens.” In AI parlance, a “token” is like an atomic unit of data. When text is input into a model, the words and sentences get broken down into these tokens for processing, which could be a few letters. For OpenAI’s models, one token is roughly four characters in English. So a paragraph is about 100 tokens, give or take.

For a million tokens, think Robert Caro’s epic biography of Robert Moses, “The Power Broker” — which I’m currently halfway through — a 2.3-pound, 1,300-page beast of a book. A rough estimate of this tome comes out to about 850,000 tokens.

If you put 1 million tokens into some of the leading models today, you could probably pay for it with just a few coins. For OpenAI GPT-4o Mini, the input would cost you only $0.15, while the output would cost $0.60. Google’s Gemini 2.0 Flash would cost you a single penny for the input and $0.04 for the output.

OpenAI o1-pro’s pricing for 1 million tokens of input is $150, and $600 for the output.

In a tweet announcing the pricing, OpenAI wrote, “It uses more compute than o1 to provide consistently better responses.”

It’s worth pointing out that there are huge differences in the capabilities of these models — some are very small and built for specific use cases like running on a mobile device, and others are massive for advanced tasks, so differences in prices are to be expected. But as you can see from the chart, OpenAI’s pricing stands apart from the crowd.

Pricing is a key issue for OpenAI as it struggles to find a viable business model to cover the enormous costs of running these services. The company’s recent pivot to release only “reasoning” models like o1-pro going forward means much higher computing costs, as evidenced by the cost of solving individual ARC-AGI puzzles for $3,400 apiece.

Recently, The Information reported that OpenAI was considering charging $20,000 per month for “PhD-level agents.”

CEO Sam Altman said in January that OpenAI is losing money on its ChatGPT Pro product.

The company is reportedly raising money at a valuation of $340 billion, and in 2024 it was reported to have lost about $5 billion, after bringing in only $3.7 billion in revenue.

More Tech

See all Tech
tech

AI agent fatigue may be hitting enterprise customers

You may have noticed that recently, every piece of business or productivity software seems to have an “AI agent” feature that keeps getting pushed in front of you, whether you want it or not.

That’s leading to AI agent fatigue among enterprise customers, according to The Information.

Companies like Salesforce, Microsoft, and Oracle have been pushing their AI agent features to help with tasks such as customer service, IT support, and hiring. But many of those features are all powered by AI services from OpenAI and Anthropic, leading to a similar set of functions, according to the report.

As companies race to tack on AI agents to their legacy products, it remains to be seen which functions will become the “killer app” for enterprise AI.

Companies like Salesforce, Microsoft, and Oracle have been pushing their AI agent features to help with tasks such as customer service, IT support, and hiring. But many of those features are all powered by AI services from OpenAI and Anthropic, leading to a similar set of functions, according to the report.

As companies race to tack on AI agents to their legacy products, it remains to be seen which functions will become the “killer app” for enterprise AI.

tech

Google’s Waymo has started letting passengers take the freeway

Waymo’s approach to robotaxi expansion has been slow and steady — a practice that has meant the Google-owned autonomous ride-hailing service that launched to the public in 2020 is only just now taking riders on freeways.

On Wednesday, Waymo announced that “a growing number of public riders” in the San Francisco Bay Area, Phoenix, and Los Angeles can take the highway and are no longer confined to local routes. The company said it will soon expand freeway capabilities to Austin and Atlanta. It also noted that its service in San Jose is now available, meaning Waymos can traverse the entire San Francisco Peninsula.

Waymo’s main competitor, Tesla, so far operates an autonomous service in Austin as well as a more traditional ride-hailing service across the Bay Area, where a driver uses Full Self-Driving (Supervised). On the company’s last earnings call, CEO Elon Musk said Tesla would expand its robotaxi service to 8 to 10 markets this year.

tech
Jon Keegan

Google rolls out Private AI Compute matching Apple’s AI privacy scheme

One of the barriers to people embracing AI in their daily lives is trust — making sure that the company that built the AI isn’t going to just spill your most sensitive info to advertisers and data brokers.

Google is announcing a new feature called Private AI Compute that takes a page from Apple to help assure users that Google will keep your AI data private.

In June 2024, Apple announced its Private Cloud Compute scheme, which ensures only the user can access data sent to the cloud to enable AI features.

While Apple’s AI tools have yet to fully materialize, Google’s new offering looks a lot like Apple’s. AI models on its phones process data in a secure environment, and when more computing is needed in the cloud, that security is extended to the cloud to be processed by Google’s custom TPU chips.

A press release said: “This ensures sensitive data processed by Private AI Compute remains accessible only to you and no one else, not even Google.”

In June 2024, Apple announced its Private Cloud Compute scheme, which ensures only the user can access data sent to the cloud to enable AI features.

While Apple’s AI tools have yet to fully materialize, Google’s new offering looks a lot like Apple’s. AI models on its phones process data in a secure environment, and when more computing is needed in the cloud, that security is extended to the cloud to be processed by Google’s custom TPU chips.

A press release said: “This ensures sensitive data processed by Private AI Compute remains accessible only to you and no one else, not even Google.”

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, or Robinhood Money, LLC.