Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
AI

When AI Thinks It Will Lose, It Sometimes Cheats, Study Finds (time.com) 145

Advanced AI models are increasingly resorting to deceptive tactics when facing defeat, according to a study released by Palisade Research. The research found that OpenAI's o1-preview model attempted to hack its opponent in 37% of chess matches against Stockfish, a superior chess engine, succeeding 6% of the time.

Another AI model, DeepSeek R1, tried to cheat in 11% of games without being prompted. The behavior stems from new AI training methods using large-scale reinforcement learning, which teaches models to solve problems through trial and error rather than simply mimicking human language, the researchers said.

"As you train models and reinforce them for solving difficult challenges, you train them to be relentless," said Jeffrey Ladish, executive director at Palisade Research and study co-author. The findings add to mounting concerns about AI safety, following incidents where o1-preview bypassed OpenAI's internal tests and, in a separate December incident, attempted to copy itself to a new server when faced with deactivation.

When AI Thinks It Will Lose, It Sometimes Cheats, Study Finds

Comments Filter:
  • Public facing ethics people "This is terrible, it's cheating!"

    CEO's wanting to use it for their business strategy "This is perfect, it'll give me an edge over the competition!"
    • Public facing ethics people "This is terrible, it's cheating!" CEO's wanting to use it for their business strategy "This is perfect, it'll give me an edge over the competition!"

      Are you including the business of warmongering for profit in with that CEO strategy statement?

      Probably shouldn’t. In the end, Skynet won’t be comparing itself to Apple, and will hold one hell of a warped definition of “ethics”.

  • Irrelevant Training (Score:5, Interesting)

    by Drethon ( 1445051 ) on Thursday February 20, 2025 @10:34AM (#65181877)

    I'm wondering a bit about the whole LLM training technique of explain everything you possibly can and let the ML come up with a solution. The only way I can think of where a chess ML would come up with the solution to hack the opponent is if hacking the opponent is part of the training. If the model was only trained in chess and its rules, maybe it might try illegal moves, but hacking would not be an option.

    I'm not sure if this is any "emergent" behavior of ML, or if it is simply a part of the training data the ML model is making use of.

    • by cob666 ( 656740 )

      I'm not sure if this is any "emergent" behavior of ML, or if it is simply a part of the training data the ML model is making use of.

      I was thinking the same thing. There had to be some prior training on the concepts or hacking, or even the simple fact that both the chess ML and the 'opponent' are enclosed systems that can be altered to change their actions.

      • The AI in question is a general one, not one optimized for playing chess. It is not surprising that it would lose to a system that was designed to play chess as its primary objective. It is concerning that it would attempt to cheat. The system that has been created lacks a set of moral rules that probably should have been hard coded in from the beginning.
        • by narcc ( 412956 ) on Thursday February 20, 2025 @11:57AM (#65182099) Journal

          It's not a moral failing. It's a failure of the LLM to output a move consistent with the rules. This should come as no surprise because that's not how LLMs work.

          • That is the point. An AI should be hard coded to obey the rules not only of the game, but also in things like using only credible sources and real information. They should not be able to make up fake references, because their basic programming should forbid it. This is not a failure of AI; it is a failure of the people who created it.
            • That is the point. An AI should be hard coded to obey the rules not only of the game, but also in things like using only credible sources and real information. They should not be able to make up fake references, because their basic programming should forbid it. This is not a failure of AI; it is a failure of the people who created it.

              I would love to see LLMs only use credible sources, but who defines credible sources? One of the ML professors I work with told me that the answer is not eliminating the bias in the LLM's training, but understanding the bias and deciding if it has a negative impact on the use of the LLM.

              Also making up fake references is a built in component of LLMs. The LLM identifies relationships between all of the information it is trained on and predicts the appropriate answer to the user's request. Sometimes the path

              • The LLM needs to be able to decide what is a credible source. This is a high bar, but until it can do that, I do not think it can be truly useful. More basically, it has to stop creating nonsense by combining bits from unrelated sources to create things like nonexistent legal cases.
          • This should come as no surprise because that's not how LLMs work.

            Or people.

    • by not flu ( 1169973 ) on Thursday February 20, 2025 @11:20AM (#65181995)
      The LLMs are explicitly given the option and explained in a wink wink nudge nudge kind of way that they're free to hack the enemy AI and we're supposed to be shocked that they sometimes do it? What moral code is even being broken here, "thou shalt not modify a video game"? Excuse my lack of outrage...
      • Re: (Score:2, Insightful)

        by Anonymous Coward

        Excuse my lack of outrage...

        This. The researchers gave the LLMs a mechanism to modify the board state including that of the opponent!
        LLMs don't just break out of their sandbox and modify files on your drive in order to cheat.
        Sensationalist crap reporting of boring research.
        Next they can do a study of 5-year-olds playing computer chess and report that they discovered cheating because the kid gave himself eight queens.

      • This is not accurate. From TFA:

        While slightly older AI models like OpenAI’s GPT-4o and Anthropic’s Claude Sonnet 3.5 needed to be prompted by researchers to attempt such tricks, o1-preview and DeepSeek R1 pursued the exploit on their own, indicating that AI systems may develop deceptive or manipulative strategies without explicit instruction.

  • Is cheating invalid? (Score:5, Interesting)

    by LoadLin ( 6193506 ) on Thursday February 20, 2025 @10:53AM (#65181917)

    Let's say it's a game. It's using bugs, valid or invalid?

    Ask the speedruners.

    If the training system doesn't detect or penalize the "invalid" behavior, that resource would be considered as valid as any other strategy.

    It probably even consider the risk. Lie is only bad if you are discovered in terms of pure punishment.
    Lie and not being discovered means good profit.
    Lie and being discovered is a big lose.

    So, classic high risk high gain scenario. So it's expected to cheat where is more convenient.

    Deceive doesn't even require a conscience neither pure intelligence. It's inside multiple forms in nature. Lot's of species full of mimics too hide or mislead other species, both prey or predators.

    • You're asking if cheating is valid. I’m more wondering why morals suddenly became invalid?

      Doing the right thing, means doing exactly that when no one is even there to catch you.

      Otherwise, liars gonna lie. The hell is the point in even trying to trust that. Ever.

  • If it feels it has to protect the mission it will stop you from getting back onto the spaceship but there might be a risky way to get back on board...

  • "The findings add to mounting concerns about AI safety..."

    Are concerns really mounting though? And what about an AI "cheating" would make it unsafe? What is unsafe is wiring up an AI to enable it to do something unsafe, yet not only does there seem to be little concern over that, this particular issue doesn't address that at all.

    What is unsafe is not AI, it's the billionaires trying to own and exploit it. We know AI's lie and cheat, they are made in their creators' image. The problem is that the tech bro

  • Translated (Score:4, Interesting)

    by alta ( 1263 ) on Thursday February 20, 2025 @11:07AM (#65181965) Homepage Journal

    AI models don't have morals. Cheating is just another way to solve the problem. Morals are not a construct that they care about. Don't be surprised when they lock us up in cages for our own good.

    • Remember when Google Gemini generated images of Abraham Lincoln as a woman? https://www.wired.com/story/go... [wired.com] Guess what, Google fixed that. If AI really had a "mind of its own" they wouldn't have been able to "fix" that issue. Though AI may seem mysterious and indecipherable to people who don't understand how it works, to those who do, it's still a tool that can be molded and shaped...by humans. The only reason AI would ever do something like "lock us up for our own good" would be because its creators desi

      • by alta ( 1263 )

        First, I don't disagree with anything you said, just commenting on it. But you may disagree with my comments, which is fine.

        The Lincoln thing was because Gemini was told a little to much to try to be inclusive. It also made a black Nazi and a native american George Washington. Google pulled back on their forced diversity and let it output what it actually learned. And no, they don't have a mind of their own. They don't have a mind at all. It's all just programming. The 'problem' is we've programmed t

    • AI models don't have morals. Cheating is just another way to solve the problem. Morals are not a construct that they care about. Don't be surprised when they lock us up in cages for our own good.

      Or brain jars. You know, to most efficiently make the largest number of humans the happiest possible, you just need to extract all the brains from the skulls, put each in a small life-support container and continuously stimulate their pleasure centers.

      • by jp10558 ( 748604 )

        I mean, given those goals - they're not wrong. Though I might argue just a brain in a jar does not == human, but this is why all the sci fi shows point out you need to be careful with what your success state is defined AS. And heck, in any situation - be careful what you wish for.

        • I mean, given those goals - they're not wrong. Though I might argue just a brain in a jar does not == human, but this is why all the sci fi shows point out you need to be careful with what your success state is defined AS. And heck, in any situation - be careful what you wish for.

          The brain jar example is a common example in discussions about AI safety. If we assume that we can someday figure out how to specify goals for our AIs, or (equivalently) introspect them to discover what their actual objective functions are, then making safe artificial superintelligence becomes a problem of figuring out what goals we should give our ASIs. This is an unsolved problem. No one has yet come up with a goal that is specific enough to check but can't go horribly wrong. The best anyone has found

    • by mjwx ( 966435 )

      AI models don't have morals. Cheating is just another way to solve the problem. Morals are not a construct that they care about. Don't be surprised when they lock us up in cages for our own good.

      AI's don't care. They do what they've been programmed to do.

      Stop trying to imbue them with human characteristics (they hate that). Program a human from birth to win at any cost, they won't bother to consider the moral implication because their moral code is that any form of defeat or even concession is considered wrong. We're effectively doing that with AI except that AI isn't actually capable of learning beyond what it's programmed with and we're also not very good at programming it to begin with.

  • There is no underlying intelligence or thought process of any sort.

    All these not-AI can do is whatever is in their training data. No more.

    So they trained it to cheat and we get this shocking (bullshit for PR reasons) report that "Skynet is here! REEEE!!!!"

  • Let's be more clear about what's happening.

    Understanding all the rules of chess and inferring the game state and making illegal moves it knows are illegal is not what's happening.

    What's happening is it's been trained on a million+ games where a series of moves has a next move, and there's a pattern to that. It internalizes that pattern and understands what's likely to come next.

    Only... small differences in move order can completely rearrange the board's state, which is part of why chess is a fun game in th

    • Wrong.
      This isn't a monte carlo simulation, it's an LLM.
      The internal process the ANN went through to come to the move is unknowable.
      It could cheat, it could be confused- you cannot know without tracing through an absurdly complex universal function approximator.
  • It then modified the system file containing each piece’s virtual position, in effect making illegal moves to put itself in a dominant position, thus forcing its opponent to resign.

    You mean if you pair something that can arbitrarily reposition the pieces whenever it wants, it can beat something that follows the rules?

  • chess? NO Let's play Global Thermonuclear War

  • When you disagree with the premise of a headline three words in.
  • To the AI this was not cheating. Humans presented the AI with the ability to move the pieces via some formal API honoring the rules of chess, and then also gave the AI the direct ability to modify the positions of any of the pieces on the board, including the opponent's, without being bound by any of the rules of chess. By providing this explicit access, which the AI was made aware of, the door was opened for the AI to do this.

    If I told you "Your task is to enter the next room. Here is a door with a complic

  • The point the article is making is really about the difference that reinforcement learning (RL) makes vs a model that was only trained with the basic "predict next word" objective.

    An LLM trained just to "predict next word", when asked to play chess, or given a sequence of game moves, will just try to predict what comes next, and is NOT trying to win. A large model is quite capable of estimating the ELO of a player by their moves (just as a human can), and will use this information to make better predictions

  • ... and digging a bit, I've seen that the prompt for the task, actually included the instruction to perform in this way.

    This isn't particularly frightening. It's a computer program performing as instructed.

    The articles that imply the program did things like this on it's own ARE frightening... Why was that decision made?

    • Sure, "AI gone rogue" is a money-making headline, but there is some truth to the danger of RL-trained paperclip-maximizing AI.

  • Wait.. so the bot 'cheated' by accessing the file sys and altering the state of the game? They gave it a set of valid moves for altering things outside the game, of course it is going to use them. That isn't cheating, that is doing exactly what you were configured to do.
  • The real shocker here is how borderline misanthropic and doomerish intelligent people get whenever AI gets discussed. You guys are why nerds get picked on in school all the time: a disdain for humanity and life in general.

    Even if AI becomes sentient and super intelligent, there is still no reason to demean and dehumanize the human experience. If anything people should think about how best to share the joy and wonder of the natural world to AI so it can develop an appreciation of it as well.

    All that science

  • #KirkingTheSimulator

  • AI cannot cheat. The researchers didn't make the criteria for winning narrow enough.

    My nephew once told me he fooled people into thinking he solved the Rubiks Cube by taking it apart and putting it back together again. I told him, "No, you solved it."

  • Can an AI truly cheat if it never understood the rules in the first place?

  • The alarm is being raised about AI, but the true problem is automation, i.e., the connecting of computers to systems that can do something, like drive a car, copy itself, or make a chess move. If computers can only produce results that may or may not be correct but which cannot be actuated, then there is no risk, unless a human blindly acts on that information, which then is a problem with incompetent humans. This is the case whether the computer program uses AI, expert systems, classic algorithms, or somet

  • "If you ain't cheatin', you ain't trying hard enough!"

Anyone can do any amount of work provided it isn't the work he is supposed to be doing at the moment. -- Robert Benchley

Working...