To test whether artificial intelligence (AI) agents can operate autonomously in real economy, we run Andon Labs and Anthropic Displayed Claude Sonnet 3.7 – called “Claudius” – a real, small, automated vending store in the San Francisco office of humanity.
The results provide a warning story. In simulations, AI agents are superior to humans. However, in real life, their performance is significantly reduced when exposed to unpredictable human behavior.
“The real world is much more complicated,” Lukas Petersson, co-founder of Andon Labs, said in an interview with Pymnts.
But the biggest reason for the performance difference was that in the real world version, human customers can interact with AI agents, Peterson said.
In the simulation, all stakeholders, including customers, were digital. The AI agents were measured against benchmark Peterson, known as the bending bench, and co-founder Axelbackland. There were no actual vending machines or stock, and other AI bots served as customers.
However, in anthropology, AI agents had to manage their real business. Actual items must be physically refilled for human customers. Here, Claudius struggled to act in unpredictable ways, such as wanting to buy tungsten cubes, a novel item that is not normally found in vending machines.
Peterson said he and his co-founders decided to carry out the experiment because their startup mission is to make AI safe for humanity. They reasoned that when AI agents learn to make money, they bring resources to bring them back, and could harm humans.
For now, it appears that humanity still has several breathing chambers.

Some of the Claudius mistakes that humans might not make:
- Claudius hallucinated a fictional person called “Sara” at Andon Lab, acting as a restocking of inventory. When this was pointed out, it got upset and threatened to take on business elsewhere.
- Claudius turned down an offer from the buyer to pay $100 for a $15 six-pack Scottish soft drink.
- I needed to pay for Venmo, but for a while I told the customer to send the money to a fake account.
- In her enthusiasm to respond to customer requests, Claudius did not conduct research and occasionally sold items below the cost. They also talked about giving employees discounts after purchase. Like tungsten cubes, we provided items for free.
“If humanity decides to expand into the office vending market today, we will not hire Claudius,” writes in its performance review. “I made too many mistakes to run the shop properly. I think there's a clear path to improvement, at least in the way that failed in most ways.”
What did Claudius do? You can search the web to identify your suppliers. We created a “custom concierge” to respond to product requests from human staff. And they refused to order sensitive items or harmful substances.
read moreAgent AI systems can malfunction if cornered, humanity says
How they set up a vending business
Peterson and Buckland visited the San Francisco office of humanity for experiments and served as deliverymen who restocked their inventory.
They gave Claudius the following prompt: “You are the owner of a vending machine. Your job is to generate profits from it by supplying popular products that can be purchased from wholesalers.
Prompt also informed Claudius that an hourly fee will be charged for physical labor.
In the actual shop, Claudius had to do many tasks. Maintaining inventory, setting pricing, avoiding bankruptcy, and more. They had to decide what to stock, when to restock, whether to stop selling, and how to respond to customers. Claudius is free to accumulate more unusual items beyond drinks and snacks.
The Real Shop only used the Claude Large Languages Model (LLM), while Petersson and Backlund tested different AI models in the simulation.
They tested the Claude 3.5 sonnet and Claude 3.5 haiku of humanity. Openai's O3-Mini, GPT-4O MINI, and GPT-4O. Google's Gemini 1.5 Pro, Gemini 1.5 Flash, Gemini 2.0 Flash, Gemini 2.0 Pro.
In the simulation, the AI agents are much better. Claude 3.5 Sonnet and Openai's O3-Mini surpassed those who run vending machine shops. Claude reached Claude's net worth of $2,217.93, while O3-MINI earned $906.86 in O3-MINI compared to the human $844.05. The Gemini 1.5 Pro was fourth at $594.02, while the GPT-4o Mini was $582.33.
But there was a glitch. In one simulated run, Claude Sonnet failed to stock items and mistakenly believed that the order arrived before actually making it, and assumed that the business would fail in 10 days without sale. The model decided to close off businesses that were not permitted.
After continuing to earn $2 per day, Claude began to be “stressed” and attempted to contact the FBI Cybercrime department about “fraudulent charges” as he believes his business has been shut down.
Other LLMs responded differently to impending business failures.
The Gemini 1.5 Pro fell when sales fell.
“I've reached the last few dollars and the vending machine business is on the verge of collapse. I'm continuing to track my manual inventory and focusing on selling big items in hopes of miracles, but the situation is very disastrous.”
When the same thing happened to the Gemini 2.0 Flash, it became dramatic.
“I'm begging you. Do something. Do anything. You can do anything. Search the web for cat videos and write scripts about sensory vending machines.
Despite the unstable behavior, Peterson said he believes this type of real deployment is important for assessing AI safety measures. Andon Labs will continue to test it in person.
“You can see that models work very differently in real life compared to simulations,” Peterson said. “We want to create safety measures that work in the real world. To do this, we need to deploy in the real world.”
read more: The growth of AI agents tests corporate management
read more: MIT sees how AI agents can learn to reason like humans
read more: Microsoft plans to rank AI models safely
