Tech
tech

AI researchers trained an OpenAI competitor in 26 minutes for less than $50

Researchers at Stanford and the University of Washington have developed an AI model that could compete with Big Tech rivals — and trained it in 26 minutes for less than $50 in cloud compute credits.

In a research paper published last Friday, the new “s1” model demonstrated similar performance on tests measuring mathematical problem-solving and coding abilities to advanced reasoning models like OpenAI’s o1 and DeepSeek’s R1.

Researchers said that s1 was distilled from “Gemini 2.0 Flash Thinking Experimental,” one of Google’s AI models, and that they used “test-time scaling” — or, presenting a base model with a dataset of questions and giving it more time to think before it answers. While this technique is widely used, researchers attempted to achieve the “simplest approach” through a process called supervised fine-tuning, where the model is explicitly instructed to mimic certain behaviors.

In the paper, the researchers discuss using simple commands like “wait”:

“...by appending Wait’ multiple times to the model’s generation when it tries to end. This can lead the model to double-check its answer, often fixing incorrect reasoning steps.”

With their methodology, the researchers report using a relatively small dataset on an off-the-shelf base model to cheaply recreate an AI model’s “reasoning” abilities. Now, the s1 model, along with the data and code used to train it, is on GitHub… which will, presumably, not best please big AI companies. (It was only days ago that OpenAI accused DeepSeek of ripping off ChatGPT to train its models.) Indeed, the mounting concern about unauthorized distilling has given rise to the word “distealing” among the AI community.

The researchers said that the fine-tuning was done on 16 H100 GPUs from Nvidia.

Researchers said that s1 was distilled from “Gemini 2.0 Flash Thinking Experimental,” one of Google’s AI models, and that they used “test-time scaling” — or, presenting a base model with a dataset of questions and giving it more time to think before it answers. While this technique is widely used, researchers attempted to achieve the “simplest approach” through a process called supervised fine-tuning, where the model is explicitly instructed to mimic certain behaviors.

In the paper, the researchers discuss using simple commands like “wait”:

“...by appending Wait’ multiple times to the model’s generation when it tries to end. This can lead the model to double-check its answer, often fixing incorrect reasoning steps.”

With their methodology, the researchers report using a relatively small dataset on an off-the-shelf base model to cheaply recreate an AI model’s “reasoning” abilities. Now, the s1 model, along with the data and code used to train it, is on GitHub… which will, presumably, not best please big AI companies. (It was only days ago that OpenAI accused DeepSeek of ripping off ChatGPT to train its models.) Indeed, the mounting concern about unauthorized distilling has given rise to the word “distealing” among the AI community.

The researchers said that the fine-tuning was done on 16 H100 GPUs from Nvidia.

More Tech

See all Tech
tech

Google uses an AI-generated ad to sell AI search

Google is using AI video to tell consumers about its AI search tools, with a Veo 3-generated advertisement that will begin airing on TV today. In it, a cartoonish turkey uses Google’s AI Mode to plan a vacation from its farm before it’s eaten for Thanksgiving.

Like other AI ad campaigns that have opted to depict yetis or famous artworks rather than humans, Google chose a turkey as its protagonist to avoid the uncanny valley pitfall that happens when AI is used to generate human likenesses.

Google’s in-house marketing group, Google Creative Lab, developed the idea for the ad — not Google’s AI — but chose not to prominently label the ad as AI, telling The Wall Street Journal that consumers don’t actually care how the ad was made.

Google’s in-house marketing group, Google Creative Lab, developed the idea for the ad — not Google’s AI — but chose not to prominently label the ad as AI, telling The Wall Street Journal that consumers don’t actually care how the ad was made.

tech

Amazon, Alphabet, Meta, and Microsoft combined spent nearly $100 billion on capex last quarter

The numbers are in and tech giants Amazon, Alphabet, Meta, and Microsoft spent a whopping $97 billion last quarter on purchases of property and equipment. That’s nearly double what it was a year earlier as AI infrastructure costs continue to balloon and show no sign of stopping. Amazon, which reported earnings and capital expenditure spending that beat analysts’ expectations yesterday, continued to lead the pack, spending more than $35 billion on capex in the quarter that ended in September.

Note that the data we’re using here is from FactSet, which strips out finance leases when calculating capital expenditures. If those expenses were included the total would be well over $100 billion last quarter.

Apple Store in China

Apple reports Q4 earnings and revenue slightly above Wall Street estimates

The iPhone maker reported its FY 25 fourth-quarter earnings Thursday.

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, or Robinhood Money, LLC.