Tech
tech

AI researchers trained an OpenAI competitor in 26 minutes for less than $50

Researchers at Stanford and the University of Washington have developed an AI model that could compete with Big Tech rivals — and trained it in 26 minutes for less than $50 in cloud compute credits.

In a research paper published last Friday, the new “s1” model demonstrated similar performance on tests measuring mathematical problem-solving and coding abilities to advanced reasoning models like OpenAI’s o1 and DeepSeek’s R1.

Researchers said that s1 was distilled from “Gemini 2.0 Flash Thinking Experimental,” one of Google’s AI models, and that they used “test-time scaling” — or, presenting a base model with a dataset of questions and giving it more time to think before it answers. While this technique is widely used, researchers attempted to achieve the “simplest approach” through a process called supervised fine-tuning, where the model is explicitly instructed to mimic certain behaviors.

In the paper, the researchers discuss using simple commands like “wait”:

“...by appending Wait’ multiple times to the model’s generation when it tries to end. This can lead the model to double-check its answer, often fixing incorrect reasoning steps.”

With their methodology, the researchers report using a relatively small dataset on an off-the-shelf base model to cheaply recreate an AI model’s “reasoning” abilities. Now, the s1 model, along with the data and code used to train it, is on GitHub… which will, presumably, not best please big AI companies. (It was only days ago that OpenAI accused DeepSeek of ripping off ChatGPT to train its models.) Indeed, the mounting concern about unauthorized distilling has given rise to the word “distealing” among the AI community.

The researchers said that the fine-tuning was done on 16 H100 GPUs from Nvidia.

Researchers said that s1 was distilled from “Gemini 2.0 Flash Thinking Experimental,” one of Google’s AI models, and that they used “test-time scaling” — or, presenting a base model with a dataset of questions and giving it more time to think before it answers. While this technique is widely used, researchers attempted to achieve the “simplest approach” through a process called supervised fine-tuning, where the model is explicitly instructed to mimic certain behaviors.

In the paper, the researchers discuss using simple commands like “wait”:

“...by appending Wait’ multiple times to the model’s generation when it tries to end. This can lead the model to double-check its answer, often fixing incorrect reasoning steps.”

With their methodology, the researchers report using a relatively small dataset on an off-the-shelf base model to cheaply recreate an AI model’s “reasoning” abilities. Now, the s1 model, along with the data and code used to train it, is on GitHub… which will, presumably, not best please big AI companies. (It was only days ago that OpenAI accused DeepSeek of ripping off ChatGPT to train its models.) Indeed, the mounting concern about unauthorized distilling has given rise to the word “distealing” among the AI community.

The researchers said that the fine-tuning was done on 16 H100 GPUs from Nvidia.

More Tech

See all Tech
tech

Alphabet’s Waymo to add 200 square miles of coverage area to existing markets

Waymo, a subsidiary of Alphabet, announced today that it’s expanding its coverage area by 200 square miles in several existing markets, including Miami, the San Francisco Bay Area, Houston, Austin, and Atlanta. That will bring its total coverage area to more than 1,400 square miles. The autonomous car service is currently offering public rides in 11 markets, after expanding to Nashville last month.

25%

AI companies are amping up their spending in Washington as they push for federal approval for more data centers and industry-friendly rules regarding their use of copyrighted material, among other asks, The New York Times reports, citing data from nonprofit watchdog Public Citizen. 25% of currently registered federal lobbyists are now involved in pushing AI interests. That’s more than double what it was — 11% — in 2023. Meta, Nvidia, and Alphabet spent $47.8 million combined last year, up 22% from 2024.

tech

Microsoft’s LinkedIn to lay off 5% of staff, Reuters reports

Reuters is reporting that Microsoft subsidiary LinkedIn is preparing to lay off 5% of its 17,500 staff, the latest in a string of tech cutbacks this year. Reuters doesn’t yet know what teams are affected, but a source said the reason wasn’t AI replacing jobs. LinkedIn sales rose 12% last quarter compared with a year earlier, representing accelerating growth.

Latest Stories

Sherwood Media, LLC and Chartr Limited produce fresh and unique perspectives on topical financial news and are fully owned subsidiaries of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, Robinhood Money, LLC, Robinhood U.K. Ltd, Robinhood Derivatives, LLC, Robinhood Gold, LLC, Robinhood Asset Management, LLC, Robinhood Credit, Inc., Robinhood Ventures DE, LLC and, where applicable, its managed investment vehicles.