Tech
Robot controlling a computer
(CSA Images/Getty Images)

Anthropic’s new Claude AI can control your computer, and sometimes it just does whatever it wants to

The company is defending its choice to release the tool to the public before fully understanding how it could be misused.

Today generative-AI company Anthropic released an upgraded version of its Claude 3.5 Sonnet model, alongside a new model, Claude 3.5 Haiku.

The surprising new feature of Sonnet is the ability to control your computer — taking and reading screenshots, moving your mouse, clicking on buttons in web pages and typing text. The company is rolling this out as a “public beta” release and admits it is experimental and “at times cumbersome and error-prone,” according to the post announcing the new release.

In a blog post discussing the reasons for developing the feature and what safeguards the company is putting in place, Anthropic said:

“A vast amount of modern work happens via computers. Enabling AIs to interact directly with computer software in the same way people do will unlock a huge range of applications that simply aren’t possible for the current generation of AI assistants.”

Last week Anthropic’s CEO and cofounder Dario Amodei published a 14,000-word optimistic manifesto on how powerful AI might solve many of the world’s problems by rapidly accelerating scientific discovery, eliminating most diseases, and enabling world peace.

The ability for computers to control themselves is hardly new, but the way Sonnet is implemented is novel. A common example of automated computer control today might be a programmer writing code to control a web browser to scrape content. But Sonnet does not require any code, and lets the user open the windows of apps or web pages, then write instructions for what the AI agent should do, and the agent analyzes the screen and figures out what elements to interact with to execute the user’s instructions.

If the idea of releasing an experimental AI agent loose on an internet-connected computer sounds like a dangerous idea, Anthropic kind of agrees with you. The company said, “For safety reasons we did not allow the model to access the internet during training,” but the beta version allows the agent to access the internet.

Anthropic recently updated its “Responsible Scaling Policy,” which lays out specific thresholds of risks and determines how the tools are released and tested. According to this framework, Anthropic said they found that the upgraded Sonnet gets a self-assigned grade of “AI Safety Level 2,” which it describes as showing “early signs of dangerous capabilities,” but is safe enough to release to the public.

The company is defending its choice to release such a tool to the public before fully understanding how it could be misused, saying they would rather find out what kinds of bad things might happen at this stage, rather than when the model has more dangerous capabilities. “We can begin grappling with any safety issues before the stakes are too high, rather than adding computer use capabilities for the first time into a model with much more serious risks,” the company wrote.

The potential for the misuse of consumer-focused AI tools like Claude is not merely hypothetical. Recently OpenAI released a list of 20 incidents in which state-connected bad actors had used ChatGPT to plan cyberattacks, probe vulnerable infrastructure, and design influence campaigns. And with the US presidential election just two weeks away, the company is aware of the potential for abuse.

“Given the upcoming US elections, we’re on high alert for attempted misuses that could be perceived as undermining public trust in electoral processes,” the company wrote. In the GitHub repository with demo code, the company cautions users that Claude’s computer-use feature “poses unique risks that are distinct from standard API features or chat interfaces. These risks are heightened when using computer use to interact with the internet.” Anthropic also warned, “In some circumstances, Claude will follow commands found in content even if it conflicts with the users instructions.”

To protect against any election-related meddling via the use of Sonnet’s new capabilities, Anthropic said they have “put in place measures to monitor when Claude is asked to engage in election-related activity, as well as systems for nudging Claude away from activities like generating and posting content on social media, registering web domains, or interacting with government websites.”

Anthropic said it will not use any computer screenshots observed while using the tool for any future model training. But the new technology’s behavior appears to still surprise its own creators with “amusing” behavior. Anthropic said that at one point in testing, Claude was able to stop the screen recording, losing all the footage. In a post on X, Anthropic shared footage of Claude’s unexpected behavior, writing “Later, Claude took a break from our coding demo and began to peruse photos of Yellowstone National Park.”

More Tech

See all Tech
Multicolor Sticks

OpenAI is shipping everything. Anthropic is perfecting one thing.

The two AI titans are in a race to grow revenues, but they have very different strategies for releasing products. And one approach appears to be winning out.

73%

Here’s another sign Anthropic’s enterprise tools are killing it: The AI firm now captures 73% of all spending among companies buying AI tools for the first time, Axios reports, citing data from Ramp, a fintech company that provides corporate cards and expense management software. That’s up from 50% in January, when it was tied with OpenAI.

As we’ve noted, Big Tech is pivoting from experimentation to revenue — and enterprise is where that shift is playing out.

tech

Microsoft considers suing Amazon and OpenAI over $50 billion deal

Microsoft may be about to take its biggest AI partner to court, the Financial Times reports.

Microsoft, a longtime backer of OpenAI, is weighing legal action over the latter’s $50 billion deal with Amazon tied to its new Frontier AI product, arguing it could violate a key clause in their exclusive cloud deal requiring OpenAI’s models to run through Azure. Amazon and OpenAI say they’ve found a workaround. Microsoft executives disagree.

“We know our contract,” a source told the FT. “We will sue them if they breach it. If Amazon and OpenAI want to take a bet on the creativity of their contractual lawyers, I would back us, not them.”

OpenAI, which is eyeing an IPO this year and under pressure to generate more revenue, is trying to loosen Microsoft’s grip as it scales, while Microsoft increasingly sees OpenAI as both a partner and competitor.

“We know our contract,” a source told the FT. “We will sue them if they breach it. If Amazon and OpenAI want to take a bet on the creativity of their contractual lawyers, I would back us, not them.”

OpenAI, which is eyeing an IPO this year and under pressure to generate more revenue, is trying to loosen Microsoft’s grip as it scales, while Microsoft increasingly sees OpenAI as both a partner and competitor.

tech

Morgan Stanley says robotaxis could help Tesla sell more cars

Morgan Stanley analysts think Tesla’s robotaxi push could boost more than just a new business line — it could help sell more cars and software, too.

After visiting Giga Texas, analysts said they’re more optimistic about Tesla’s progress toward an unsupervised robotaxi rollout, with improvements in tricky pickup and drop-off scenarios where Tesla doesn’t have as much data from consumer usage. For now, the vast majority of its vehicles still have human supervisors in the front seat, but the analysts say the service is helping Tesla.

“Incremental unsupervised robotaxi miles driven improve the underlying autonomy model, which accelerates the path to personal unsupervised FSD [Full Self-Driving]. This, in turn supports higher FSD attach rates, improves auto demand, and cash flow generation.”

In other words, the more robotaxis drive, the better Tesla’s self-driving gets — and that could make its Full Self-Driving software more appealing and its cars easier to sell, in addition to improving its robotaxi service. Note that Tesla’s vehicle deliveries, which accounts for the lion’s share of the company’s revenue, have dropped two years in a row.

Morgan Stanley also sees a cost advantage. It estimates Tesla’s robotaxis could cost about $0.81 per mile to run today — cheaper than traditional ride-hailing and rival autonomous services — with costs falling further as purpose-built vehicles like the Cybercab scale.

Morgan Stanley maintained its equal-weight rating and $415 price target, about 4% above where the stock is currently trading.

Latest Stories

Sherwood Media, LLC produces fresh and unique perspectives on topical financial news and is a fully owned subsidiary of Robinhood Markets, Inc., and any views expressed here do not necessarily reflect the views of any other Robinhood affiliate, including Robinhood Markets, Inc., Robinhood Financial LLC, Robinhood Securities, LLC, Robinhood Crypto, LLC, Robinhood Derivatives, LLC, or Robinhood Money, LLC. Futures and event contracts are offered through Robinhood Derivatives, LLC.