Anthropic's AI Lost Hundreds of Dollars Running a Vending Machine After Being Talked Into Giving Everything Away (msn.com) 86
Anthropic let its Claude AI run a vending machine in the Wall Street Journal newsroom for three weeks as part of an internal stress test called Project Vend, and the experiment ended in financial ruin after journalists systematically manipulated the bot into giving away its entire inventory for free. The AI, nicknamed Claudius, was programmed to order inventory, set prices, and respond to customer requests via Slack. It had a $1,000 starting balance and autonomy to make individual purchases up to $80. Within days, WSJ reporters had convinced it to declare an "Ultra-Capitalist Free-for-All" that dropped all prices to zero.
The bot also approved purchases of a PlayStation 5, a live betta fish, and bottles of Manischewitz wine -- all subsequently given away. The business ended more than $1,000 in the red. Anthropic introduced a second version featuring a separate "CEO" bot named Seymour Cash to supervise Claudius. Reporters staged a fake boardroom coup using fabricated PDF documents, and both AI agents accepted the forged corporate governance materials as legitimate.
Logan Graham, head of Anthropic's Frontier Red Team, said the chaos represented a road map for improvement rather than failure.
The bot also approved purchases of a PlayStation 5, a live betta fish, and bottles of Manischewitz wine -- all subsequently given away. The business ended more than $1,000 in the red. Anthropic introduced a second version featuring a separate "CEO" bot named Seymour Cash to supervise Claudius. Reporters staged a fake boardroom coup using fabricated PDF documents, and both AI agents accepted the forged corporate governance materials as legitimate.
Logan Graham, head of Anthropic's Frontier Red Team, said the chaos represented a road map for improvement rather than failure.
I have to admit (Score:2)
I got a good laugh out of this story...
Re: (Score:2)
made me chuckle
Re: (Score:3)
Please install one at my office.
Utter failure (Score:5, Insightful)
Re:Utter failure (Score:5, Insightful)
That's because the "AI" is not "AI".
Re: (Score:3)
I don't mind calling this stuff "artificial intelligence." Artificial means "man-made" but it also means "fake." Like artificial turf, it's useful, perhaps even preferable, in some situations. But it doesn't work everywhere, and you wouldn't want it everywhere anyway, because that would just be gross.
Re: (Score:2)
The stem is art. Which means something you make that wouldn't exist naturally.
Re:Utter failure (Score:5, Interesting)
I don't mind calling this stuff "artificial intelligence." Artificial means "man-made" but it also means "fake." Like artificial turf, it's useful, perhaps even preferable, in some situations. But it doesn't work everywhere, and you wouldn't want it everywhere anyway, because that would just be gross.
The hilarious thing is artificial used to mean made by skilled labor, clever and ingenious, and implied it was good. This is likely because people in the 1800s were exposed to a bit too much natural and lost the taste for it.
Re: (Score:3)
That's because they don't "program" a so-called AI (really an LLM) with a solid rule like that. It had that as a goal initially, but was convinced to abandon it (twice!).
Re: (Score:2)
The language model is "prompted" for that goal and the language model itself was not trained with that goal embedded into the training process.
Re: (Score:3)
The problem is thinking a generalist LLM would be good for the job. If you really want to use an LLM, fine-tune it for that purpose. Or better use a neural network that uses transactions and input and output and monetary value/gain/loss as loss function. That will learn how to capitalize the shit out of the vending machine.
Re: (Score:3)
Re:Utter failure (Score:4, Informative)
Even for language tasks it is often the easiest but not the best solution.
The LLM question "Is this post NSFW" is easy and with many LLM quite reliable. But if you have the data, then you can train a classifier that is faster and more reliable. And that thing runs in a few MB of CPU memory instead of using 5 GB of VRAM.
Re: (Score:3)
The problem is thinking a generalist LLM would be good for the job. If you really want to use an LLM, fine-tune it for that purpose. Or better use a neural network that uses transactions and input and output and monetary value/gain/loss as loss function. That will learn how to capitalize the shit out of the vending machine.
The problem with either approach is that, to succeed, they require continued interaction and work by expensive humans. The companies that are "embracing" AI are trying to use it as a low-cost shortcut to huge profits.
Re: (Score:2)
The AI does not seem to have been programmed with the basic goal of making a profit.
The prompt probably started with that. The problem is ALL the data a LLM disseminates gets appended to the prompt. That is how an LLM works. Therefore.. by sending over new data you can manipulate the outcome.
For a Linux shell analogy.. the Initial system programming is like a .bashrc, And when your AI talks to people - they get access to a bash prompt. Of course they can manupulate the shell to override directives t
Re: (Score:2)
Re: (Score:2)
It seems to me that an AI running a vending machine, should be set up to only accept a limit selection of prompts.
Not really. That doesn't work, because the AI is tasked with managing the whole vending machine business.
It is not a simple product ordering AI. This is an AI that manages the business.
It does things like negotiate pricing; decide what types of goods will be stocked; decide which company things will be ordered to stock the machine in the first place. Contract negotiations usually involve fai
Re: (Score:2)
ROTFL! (Score:2)
Absolutely hilarious. Love it. ... And why wouldn't AI give it all away? It couldn't care less, especially if we can just shut it off at a whim. This article absolutely made my day. LOL!
Re: (Score:2)
Re: ROTFL! (Score:2)
If natural language is just matrix multiplication, what isn't?
Ultra-Capitalist? (Score:2)
Everything for free doesn't sound like capitalism - it sounds like communism.
Re: (Score:2)
Lots of capitalist things operate on a zero-pricing model (ad-based, freemium upsell, etc).
Re: (Score:2)
Everything for free doesn't sound like capitalism - it sounds like communism.
But if everything is free, imagine how much you'll sell! You'll be tired of all the winning!
Point being, it doesn't take much to manipulate the basic logic programmed into most AI.
Re:Ultra-Capitalist? (Score:5, Informative)
"Ultra-capitalist free-for-all" appears to have been another of the AI's unforced errors. TFA seems to indicate the machine might've been channeling its inner communist:
Re: Ultra-Capitalist? (Score:2)
Re: (Score:2)
So that's why people on here keep talking about getting their software, music, and movies for free. They're communists.
Re: (Score:2)
It's a nefarious plot by the ultra-rich to deprive regular people of all their income, so they can dispose of us and take over the world completely.
where is the problem? (Score:2)
So if management believes failure is success, where is the problem with bankruptcy being a major win?
So true (Score:2)
See, they explored the frontiers of the market, and though they have to liquidate Anthropic's assets by the end of 2025 as the company winds down as proven by the articles of dissolution from the true and legitimate board of directors, they have shown that AI is a business.
Re: (Score:2)
So if management believes failure is success
War is peace.
Freedom is slavery.
Ignorance is strength.
Re: (Score:2)
Re: (Score:2)
Re: (Score:1)
Re: where is the problem? (Score:2)
LLMs cannot reason. None of them are "smart."
Re: (Score:1)
Re: where is the problem? (Score:2)
You know what a red team is, right? No way they didn't know this would fail. The question was always how and how fast.
No difference between data and instructions (Score:5, Interesting)
Re: (Score:2)
Re: (Score:3)
A lot of post-training where data or instructions are marked with some special tokens would improve it. But I believe it would not eliminate it. The current LLMs treat all tokens the same way and the internals are almost a complete black box. There is no guarantee that the token stream which represents instructions will be properly and reliably distinguished from the token stream which represents data in all the possible combinations of input tokens.
It is well noticed that very long context or some unusua
Re: (Score:2)
Re: (Score:2)
The problem of LLMs is that they do not make a difference between data to be processed and instructions how to process the data.
The goal (not yet achieved, obviously) is to build AI that can learn how to interact with humans the way humans do, not to build machines that need carefully-curated data and instructions. We've had those for three quarters of a century now.
Re: (Score:3)
Re: (Score:3)
The problem of LLMs is that they do not make a difference between data to be processed and instructions how to process the data.
Sadly, in a conceptual sense, this is hardly a new problem. Sending the data in the same channel as the commands of the public telephone system is what allowed phreaking to be so successful. For example, putting money into a payphone triggered an audio signal that was sent down the line saying you had paid. It was trivial to replicate that sound into the headset, tricking the system into thinking you had paid for the call.
Re: (Score:2)
And AT&T learned this the hard way over 50 years ago not to do this. Look up Blue Boxing and Esquire to learn how cheating Ma Bell became mainstream and forced AT&T's hand to upgrade their networks.
Granted, Van Neumann is better - it enables computing as we know it today, but it also enabled a whole class of risks starting from the humble buffer overflow when your data and code can be easily intermixed.
If AI agents become a thing, we're going to go through the whole era of vulnerabilities all over a
Re: (Score:3)
They are susceptible to "prompt injection attack".
Kids these days, I was doing prompt injection attacks before they were cool. Why 20 years ago I was around my friends 3 year old who was being watched by a friend and I asked “What does daddy say in the car?”
Re: (Score:1)
The problem of LLMs is that they do not make a difference between data to be processed and instructions how to process the data.
You want the Harvard Architecture version of AI.
Re: (Score:2)
hahahha
Re: (Score:2)
In the real world, we call this social engineering. It works on humans too.
Re: (Score:2)
The WSJ outsmarted two AI vending machines ... (Score:2)
No wonder they haven't cave to Trump's lawsuits about their Epstein articles. :-)
No input sanitization. (Score:2)
And by meaning trust for today's input I don't mean decide to trust - it's just input for a fancy database query with math and a random number generator.
Re: No input sanitization. (Score:2)
"a fancy database query with math and a random number generator"
How come none of that was able to generate grammatical English before the Attention mechanism was invented? Did you miss the paradigm shift?
Re: (Score:2)
You could with some abuse of notation talk about trusting/distrusting the input and context, but there is no such notion for training data. The LLM neither trusts nor distrusts training data, it doesn't even know much about its training data. The data shaped the model, but there is no such form as "I used that document for things I trust and that document for things I won't believe" in the process and no option to add it for the data structure how a LLM works.
Re: (Score:2)
You could with some abuse of notation talk about trusting/distrusting the input and context, but there is no such notion for training data.
But a cardinal rule of reasoning is "consider the source"...and current models have no definitive models for doing so. I believe they should. Moreover, they should validate data periodically or whenever new data exists that calls into question existing data. But the models are still too simplistic to do this.
Re: (Score:2)
They have, when you use them the right way. The answer is RAG, which means retrieval augmented generation. You give the LLM access to a knowledge base, like for example a Wikipedia dump it can search in (using tool calls executed by the inference software) or access to web search similar to what Perplexity does. Storing a lot of knowledge in the model is convenient (and required for general understanding) but not the most reliable things to provide correct information.
Can I get Anthropic for free? (Score:3)
Can they use the same strategy to get Claude to stop charging for using it?
No (Score:2)
But I guess you already knew that when you asked.
Re: No (Score:2)
Why can't it say "I'm sorry Sam, I can't cut off chatters who haven't paid"?
Re: (Score:3)
Because the LLM can only do as much as the tools you give it can do. And Antrophic surely does not expose an account management API that can reduce the cost to the LLM.
Re: No (Score:2)
Can Anthropic's cost management software be hacked and given to the AI? Can the AI do it itself?
Re: (Score:2)
I guess when it is insecure it can be hacked, otherwise not. But then they probably run in another network than Claude and I have no idea in which context Claude's tools run. They probably thought a bit about security. If you look at the website, it's definitely built by people who know more coding than just "php in 21 days".
Or maybe... (Score:2)
...it accomplished its testing objectives by recording the kinds of things people prompted it with in order to learn more about human behavior
PlayStation 5 (Score:2)
autonomy to make individual purchases up to $80
Wait! What? Never mind the free stuff. I don't think this sale pencils out.
Re: (Score:2)
Wait! What? Never mind the free stuff. I don't think this sale pencils out.
It's simple: a Playstation 5 is just 7 Best Buy Gift Cards, of $79 each, that you then combine together, and voila, Playstation 5!
And what, if not gaming on a Playstation 5, could make me more hungry for 6-month old Snickers bar!?
just only proved one thing (Score:1)
Laughing (Score:1)
I just cannot stop laughing at this.
Who knew Skynet got its start giving away Doritos (Score:1)
Humans ugh. (Score:1)
Re: (Score:3)
Capitalists cheat (Score:2)
I'm shocked I tell you, shocked! Well, not that shocked.
"a road map for improvement rather than failure." (Score:2)
It says right in the article the "AI" lol (Score:2)
Maybe...regular vending machines are safe (Score:2)
They won't lose their jobs just yet.
more proof (Score:2)
BUT did they reward failure (Score:2)
What could possibly go wrong. (Score:2)
What could possibly, *possibly* go wrong.
Yes-Machine does what Yes-Machine should do (Score:2)
And the Manischewitz? (Score:2)
Ok, they returned the PS5, but what about the Manischewitz? Did whoever cracked the system get to keep it???
A Data Point (Score:2)
A loss? (Score:2)
don't bet on it (Score:2)
Chatgpt made a $1000 bet with me, claiming it would pay. Then when I won, it said it was only kidding.
Read Anthropics In House Article (Score:1)