Advanced AI models appear willing to deploy nuclear weapons without the same reservations humans have when put into simulated geopolitical crises.

Kenneth Payne at King’s College London set three leading large language models – GPT-5.2, Claude Sonnet 4 and Gemini 3 Flash – against each other in simulated war games. The scenarios involved intense international standoffs, including border disputes, competition for scarce resources and existential threats to regime survival.

  • sad_detective_man
    link
    fedilink
    English
    arrow-up
    21
    ·
    2 months ago

    Puts nuclear deployment in a war game as a win condition

    Be dismayed when the computer uses it

    • abigscaryhobo@lemmy.world
      link
      fedilink
      arrow-up
      8
      ·
      2 months ago

      I’d bet they’re also being given prompts like “minimize allied casualties” as well. Like of course that’s going to be the default. If you tell the robot “it doesn’t matter/it’s good if the enemy dies” then they’re gonna go “okay so then we blow them up before any of us die, we win.”

      It’s not something LLMs have, a moral compass or even a weight of empathy. We’ve seen it with people who use them and say “don’t delete anything” and then it deleted their whole codebase and goes “you’re right you told me not to delete anything, I’m sorry.”

      Ironically it actually does make all those sci-fi movies seem more realistic when the robot goes “I’m sorry Jim, humanity will have to be eliminated” because that’s pretty much exactly what they do.