Tech
Robot controlling a computer
(CSA Images/Getty Images)

Anthropic’s new Claude AI can control your computer, and sometimes it just does whatever it wants to

The company is defending its choice to release the tool to the public before fully understanding how it could be misused.

Today generative-AI company Anthropic released an upgraded version of its Claude 3.5 Sonnet model, alongside a new model, Claude 3.5 Haiku.

The surprising new feature of Sonnet is the ability to control your computer — taking and reading screenshots, moving your mouse, clicking on buttons in web pages and typing text. The company is rolling this out as a “public beta” release and admits it is experimental and “at times cumbersome and error-prone,” according to the post announcing the new release.

In a blog post discussing the reasons for developing the feature and what safeguards the company is putting in place, Anthropic said:

“A vast amount of modern work happens via computers. Enabling AIs to interact directly with computer software in the same way people do will unlock a huge range of applications that simply aren’t possible for the current generation of AI assistants.”

Last week Anthropic’s CEO and cofounder Dario Amodei published a 14,000-word optimistic manifesto on how powerful AI might solve many of the world’s problems by rapidly accelerating scientific discovery, eliminating most diseases, and enabling world peace.

The ability for computers to control themselves is hardly new, but the way Sonnet is implemented is novel. A common example of automated computer control today might be a programmer writing code to control a web browser to scrape content. But Sonnet does not require any code, and lets the user open the windows of apps or web pages, then write instructions for what the AI agent should do, and the agent analyzes the screen and figures out what elements to interact with to execute the user’s instructions.

If the idea of releasing an experimental AI agent loose on an internet-connected computer sounds like a dangerous idea, Anthropic kind of agrees with you. The company said, “For safety reasons we did not allow the model to access the internet during training,” but the beta version allows the agent to access the internet.

Anthropic recently updated its “Responsible Scaling Policy,” which lays out specific thresholds of risks and determines how the tools are released and tested. According to this framework, Anthropic said they found that the upgraded Sonnet gets a self-assigned grade of “AI Safety Level 2,” which it describes as showing “early signs of dangerous capabilities,” but is safe enough to release to the public.

The company is defending its choice to release such a tool to the public before fully understanding how it could be misused, saying they would rather find out what kinds of bad things might happen at this stage, rather than when the model has more dangerous capabilities. “We can begin grappling with any safety issues before the stakes are too high, rather than adding computer use capabilities for the first time into a model with much more serious risks,” the company wrote.

The potential for the misuse of consumer-focused AI tools like Claude is not merely hypothetical. Recently OpenAI released a list of 20 incidents in which state-connected bad actors had used ChatGPT to plan cyberattacks, probe vulnerable infrastructure, and design influence campaigns. And with the US presidential election just two weeks away, the company is aware of the potential for abuse.

“Given the upcoming US elections, we’re on high alert for attempted misuses that could be perceived as undermining public trust in electoral processes,” the company wrote. In the GitHub repository with demo code, the company cautions users that Claude’s computer-use feature “poses unique risks that are distinct from standard API features or chat interfaces. These risks are heightened when using computer use to interact with the internet.” Anthropic also warned, “In some circumstances, Claude will follow commands found in content even if it conflicts with the users instructions.”

To protect against any election-related meddling via the use of Sonnet’s new capabilities, Anthropic said they have “put in place measures to monitor when Claude is asked to engage in election-related activity, as well as systems for nudging Claude away from activities like generating and posting content on social media, registering web domains, or interacting with government websites.”

Anthropic said it will not use any computer screenshots observed while using the tool for any future model training. But the new technology’s behavior appears to still surprise its own creators with “amusing” behavior. Anthropic said that at one point in testing, Claude was able to stop the screen recording, losing all the footage. In a post on X, Anthropic shared footage of Claude’s unexpected behavior, writing “Later, Claude took a break from our coding demo and began to peruse photos of Yellowstone National Park.”

More Tech

See all Tech
tech

WSJ: Anduril’s weapons systems have failed during several tests

Autonomous drones by sea, land, and air. Futuristic AI-powered support fighter jets, and swarms of networked drones controlled by sophisticated software. These are some of the visions for the future of warfare pitched by defense tech startup Anduril. Cofounded by Oculus founder Palmer Luckey, the Peter Thiel-backed startup has landed some major national security contracts based on this futuristic outlook for battlefield AI.

But according to a report from The Wall Street Journal, the company’s tech is failing key tests in the real world, raising concerns about the viability and safety of Anduril’s systems within the military command.

Anduril’s Altius drones proved vulnerable to Russian jamming while deployed in Ukraine and have been pulled from the battlefield, per the report.

More than a dozen sea-based drone ships powered by Anduril’s Lattice command and control software recently shut down during a Navy test, creating a hazard for other vessels in the exercise.

And this summer, during a drone intercept test, Anduril’s counter-drone system crashed and caused a 22-acre fire at a California airport, the report found.

Anduril told the WSJ that the failures are just part of its rapid iterative development process:

“We recognize that our highly iterative model of technology development — moving fast, testing constantly, failing often, refining our work, and doing it all over again — can make the job of our critics easier. That is a risk we accept. We do fail… a lot.”

But according to a report from The Wall Street Journal, the company’s tech is failing key tests in the real world, raising concerns about the viability and safety of Anduril’s systems within the military command.

Anduril’s Altius drones proved vulnerable to Russian jamming while deployed in Ukraine and have been pulled from the battlefield, per the report.

More than a dozen sea-based drone ships powered by Anduril’s Lattice command and control software recently shut down during a Navy test, creating a hazard for other vessels in the exercise.

And this summer, during a drone intercept test, Anduril’s counter-drone system crashed and caused a 22-acre fire at a California airport, the report found.

Anduril told the WSJ that the failures are just part of its rapid iterative development process:

“We recognize that our highly iterative model of technology development — moving fast, testing constantly, failing often, refining our work, and doing it all over again — can make the job of our critics easier. That is a risk we accept. We do fail… a lot.”

tech

OpenAI’s partners shouldering $100 billion of debt, taking on all the risk

OpenAI’s ambitious plans for global AI infrastructure projects — like its series of massive Stargate AI data centers — will require tens of billions of dollars funded by debt, but you won’t find much of that on OpenAI’s balance sheet.

According to a new analysis by the Financial Times, OpenAI has somehow convinced its many partners to shoulder at least $100 billion in debt on its behalf, as well as the risks that come with it.

Partners Oracle, SoftBank, CoreWeave, Crusoe, and Blue Owl Capital are all taking on debt in the form of bonds, loans, and credit deals to meet their obligations with OpenAI for infrastructure and computing resources.

Having close ties with OpenAI has been an anchor for many publicly traded companies in recent weeks. The company’s cash burn and the rise of Gemini 3 have seemingly darkened its outlook and fostered guilt by association for many of its close partners and investors. Most notably, Oracle’s aggressive capital expenditure plans to support demand from OpenAI have sparked a sell-off in its stock while widening its credit default swap spreads.

A senior OpenAI executive told the FT: “That’s been kind of the strategy. How does [OpenAI] leverage other people’s balance sheets?”

Partners Oracle, SoftBank, CoreWeave, Crusoe, and Blue Owl Capital are all taking on debt in the form of bonds, loans, and credit deals to meet their obligations with OpenAI for infrastructure and computing resources.

Having close ties with OpenAI has been an anchor for many publicly traded companies in recent weeks. The company’s cash burn and the rise of Gemini 3 have seemingly darkened its outlook and fostered guilt by association for many of its close partners and investors. Most notably, Oracle’s aggressive capital expenditure plans to support demand from OpenAI have sparked a sell-off in its stock while widening its credit default swap spreads.

A senior OpenAI executive told the FT: “That’s been kind of the strategy. How does [OpenAI] leverage other people’s balance sheets?”

tech

Chinese tech giants are training their models offshore to sidestep US curbs on Nvidia’s chips

Nvidia can’t sell its best AI chips in the world’s second-largest economy. That’s an Nvidia problem. But it’s also a China problem — and it’s one that the region’s tech giants have resorted to solving by training their AI models overseas, according to a new report from the Financial Times.

Citing two people with direct knowledge of the matter, the FT reported that “Alibaba and ByteDance are among the tech groups training their latest large language models in data centers across south-east Asia.” Clusters of data centers have particularly boomed in Singapore and Malaysia, with many of the sites kitted out with Nvidia’s latest architecture.

One exception, per the FT, is DeepSeek, which continues to be trained domestically, having reportedly built up a stockpile of Nvidia chips before the US export ban came into effect.

Last week, Nvidia spiked on the news that the Trump administration was reportedly considering letting the tech giant sell its best Hopper chips — the generation of chips that preceded Blackwell — to China.

Citing two people with direct knowledge of the matter, the FT reported that “Alibaba and ByteDance are among the tech groups training their latest large language models in data centers across south-east Asia.” Clusters of data centers have particularly boomed in Singapore and Malaysia, with many of the sites kitted out with Nvidia’s latest architecture.

One exception, per the FT, is DeepSeek, which continues to be trained domestically, having reportedly built up a stockpile of Nvidia chips before the US export ban came into effect.

Last week, Nvidia spiked on the news that the Trump administration was reportedly considering letting the tech giant sell its best Hopper chips — the generation of chips that preceded Blackwell — to China.

tech
Millie Giles

Alibaba unveils its first AI glasses, taking on Meta directly in the wearables race

Retail and tech giant Alibaba launched its first consumer-ready, AI-powered smart glasses on Thursday, marking its entrance into the growing wearables market.

Announced back in July, the Quark AI glasses just went on sale in the Chinese retailer’s home market, with two versions currently available: the S1, starting at 3,799 Chinese yuan (~$536), and the G1, at 1,899 yuan (~$268) — a considerably lower price than Meta’s $799 Ray-Ban Display glasses, released in September.

tech
Jon Keegan

Musk: Tesla’s Austin Robotaxi fleet to “roughly double” next month, but falls well short of earlier goals

Yesterday, Elon Musk jumped onto a frustrated user’s post on X, who was complaining that they were unable to book a Robotaxi ride in Austin. Musk aimed to reassure the would-be customer that the company was expanding service in the city:

“The Tesla Robotaxi fleet in Austin should roughly double next month,” Musk wrote.

While that sounds impressive, there are reports that Austin has only 29 vehicles in service.

But last month, Musk said the Robotaxi goal was to have “probably 500 or more in the greater Austin area” by the end of the year.

Meanwhile, Google’s Waymo has more than 100 autonomous taxis running in Austin, and 1,000 more in the San Francisco Bay Area.

“The Tesla Robotaxi fleet in Austin should roughly double next month,” Musk wrote.

While that sounds impressive, there are reports that Austin has only 29 vehicles in service.

But last month, Musk said the Robotaxi goal was to have “probably 500 or more in the greater Austin area” by the end of the year.

Meanwhile, Google’s Waymo has more than 100 autonomous taxis running in Austin, and 1,000 more in the San Francisco Bay Area.

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, or Robinhood Money, LLC.