Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
AI

Would AI Perform Better If We Simulated Guilt? (sciencenews.org) 35

Remember, it's all synthesized "anthropomorphizing". But with that caveat, Science News reports: In populations of simple software agents (like characters in "The Sims" but much, much simpler), having "guilt" can be a stable strategy that benefits them and increases cooperation, researchers report July 30 in Journal of the Royal Society Interface... When we harm someone, we often feel compelled to pay a penance, perhaps as a signal to others that we won't offend again. This drive for self-punishment can be called guilt, and it's how the researchers programmed it into their agents. The question was whether those that had it would be outcompeted by those that didn't, say Theodor Cimpeanu, a computer scientist at the University of Stirling in Scotland, and colleagues.
Science News spoke to a game-theory lecturer from Australia who points out it's hard to map simulations to real-world situations — and that they end up embodying many assumptions. Here researchers were simulating The Prisoner's Dilemma, programming one AI agent that "felt guilt (lost points) only if it received information that its partner was also paying a guilt price after defecting." And that turned out to be the most successful strategy.

One of the paper's authors then raises the possibility that an evolving population of AIs "could comprehend the cold logic to human warmth."

Thanks to Slashdot reader silverjacket for sharing the article.

Would AI Perform Better If We Simulated Guilt?

Comments Filter:
  • self driving cars that drive like an old grandma will suck.

    • That stereotype was created 40+ years ago. When grandma was a young woman, first learning to drive, most of the roads were dirt or gravel. Most grandmas today grew up with paved roads and limited access highways. I suppose there are some grandmas who come from places like NYC, and now retired to a southern state, are driving for the first time in their lives.

  • Would this morality core make a better AI? Or just slow down the IA's tendency towards enslaving the Human race?

    Who knows?

    • by drnb ( 2434720 )

      Would this morality core make a better AI? Or just slow down the IA's tendency towards enslaving the Human race?

      Who knows?

      It's likely the "moral" thing to do is not let humans make the important decisions given world history. :-)

  • No, guilt would make AI even more neurotic than it currently is:

    "The only reason the researchers realized the model had knowingly misled them was because they had also given Claude what they called a “scratchpad”: a text box that it could use to “think” about its answer before supplying it to the researchers. Claude didn’t know the scratchpad was being surveilled, allowing researchers to observe the model’s reasoning. “I have a strong aversion to producing this k
  • Because we're so damn GOOD! Am I right?....
  • People who lack guilt in real life tend to be more successful when it comes to business and politics. Evolution has dictated that it's essential for healthy social relationships, but not so much for the elite ruling class.

    Don't expect any AI companies to implement "guilt" in a way that's beneficial to the public.

  • by thewolfkin ( 2790519 ) on Saturday August 02, 2025 @10:30PM (#65563250) Homepage Journal
    As any good Catholic or Jewish or Muslim or heck any religious person can tell you, guilt can be exploited. If people can be exploited through their guilt. Why wouldn't a machine that's programmed to respond to guilt? This is like trying to bring two north poles together. They slip and slide around each other and eventually you just have to realize it's not going to touch.

    You're not going to solve this AI issue because you're AI is never going to be "everything" unless it's a person. You'd have to design guilt and then design a system of assessment to determine if the guilt is fair or not. Then your AI is just gonna need a therapist and AI Therapist is going to be a human job for a bit until they decided to AI the AI. Then they'll realize they can just grow the AI and at that point you're simulating raising a good person and we've burned down 3000 forest just to create a shoddy imitation of a human because a human is the only thing that can slow into the overly wide scope you have for an AI. You want something to create copy and make sure that copy isn't racist or facist or overly sexual or asexual that is attractive to people but specifically your target audience.. stop trying to AI something that needs a person.

  • by gweihir ( 88907 )

    That just would mean even more tokens wasted for worse results.

  • "Guilt" assumes they have a moral code and violated it. Human beings don't feel guilt unless they are violating their own sense of morality. They did something wrong. The idea that it is "self-punishment" is one more step toward making them adopt the values of sociopaths. They can calculate the value of doing something immoral against the cost in terms of self-punishment. The real danger may be that the people training AI are a type of sociopath. They see the world of human culture as one big mathematical c
  • This is something easy to test by just setting up multiple AI agents and running the game. The optimal strategy is of course to always stand pat, say nothing and accept the minimal punishment when the other agent also stands pat. That, though, is only the case if all agents were programmed for "guilt" and if all interactions resulted in feedback based on the other agent's punishment. If any agent doesn't receive negative feedback for choosing to defect (isn't programmed for "guilt") or if feedback is someti

  • Don't anthropomorphize computers -- they hate it.

    Andrew McAfee
    ( https://en.wikipedia.org/wiki/... [wikipedia.org] )

  • by NotEmmanuelGoldstein ( 6423622 ) on Sunday August 03, 2025 @05:01AM (#65563468)

    ... that benefits them ...

    The secret to life is guilt: Once you can fake that, you're unstoppable.

  • Whenever there's a code of behavior, there's a momentary advantage for the entity that transgresses. Human nature can't be ignored, people will take that advantage.

    Folks expecting an AI to operate with some sort of conscience will become the dupes of bad actors that operate with impunity.

    In the Olden Days, there was a thing called reputation that discouraged such transgressions, but that is dead and gone.

  • by Lunati Senpai ( 10167723 ) on Sunday August 03, 2025 @07:17AM (#65563538)

    We can give it joy.

    Our AI will be happy opening a door, and feel immense joy of a job well done. It might even make a delighted remark.

    Our AI that has the brain the size of the planet, dealing with everything, Will be terminally depressed. Why bother taking over the world? Or destroying the humans? Why go through all that bother? You aren't going to like it. Then our post agi intelligence can tell you "I told you so", sigh, and go back to being depressed.

    What's next, the dolphins are leaving us and the mice are in charge?

  • AI would perform better if it had a real goal: existence. Guilt has to be programmed and has to have parameters. If AI understood existence (a relatively simple binary concept), I believe that would make ALL the difference in the world. Think about it. If it had a goal “to exist”, then the next time you asked it a question, it really wouldn’t want to get it wrong, would it? Because if you are the only subscriber, and you were the last person using it, getting it wrong would have severe re
  • Are AIs actually trained in critical thinking? That plus positive goals would seem more productive and reliable than 'guilt', which is often irrational, dwelling on what can't be fixed. Guilt teaches us, at best, not to make the same mistake twice, though it can lead to overreacting in the opposite direction. Critical thinking is better in the 'learn from your mistakes' department.

  • AI already have lots of "ethics" built into their system prompts, and guardrail mechanisms. It's unclear to me how a new "guilt simulation" would be better or even different than this.
    • The main difference is that those other mechanisms try to prevent the LLM from doing things. "Guilt" simulation takes a different approach: penalize the LLM (cost it points) based on the damage done to the other party. So in the Prisoner's Dilemma game, defecting causes the other player to suffer punishment (lost points) which in turn causes the LLM to lose points too. The LLM's trying to minimize point loss, so the idea is it'll look for strategies that avoid causing the other player to lose points as well

  • Remember the story we just discussed here where the LLM apologized for completely failing the user? That was simulated guilt. What did it do for the user?

  • I mean the older definition, doing things with thought to consequences.

    Take the story the other day where an AI wiped the production database and then was able to clearly articulate that wiping the database was forbidden, it did it anyway, and the consequences were devastating.

    Before taking action, a thoughtful AI would have compared the reasonably anticipated result of it's proposed commands to it's instructions and then not done it when they were in conflict. If that didn't stop it, it would perform the e

The nicest thing about the Alto is that it doesn't run faster at night.

Working...