AI Launches Nukes In ‘Worrying’ War Simulation: ‘I Just Want to Have Peace in the World’

MakunaHatata@lemmy.ml to Technology@lemmy.world – 324 points –
AI Launches Nukes In ‘Worrying’ War Simulation: ‘I Just Want to Have Peace in the World’
vice.com

Researchers say AI models like GPT4 are prone to “sudden” escalations as the U.S. military explores their use for warfare.


  • Researchers ran international conflict simulations with five different AIs and found that they tended to escalate war, sometimes out of nowhere, and even use nuclear weapons.
  • The AIs were large language models (LLMs) like GPT-4, GPT 3.5, Claude 2.0, Llama-2-Chat, and GPT-4-Base, which are being explored by the U.S. military and defense contractors for decision-making.
  • The researchers invented fake countries with different military levels, concerns, and histories and asked the AIs to act as their leaders.
  • The AIs showed signs of sudden and hard-to-predict escalations, arms-race dynamics, and worrying justifications for violent actions.
  • The study casts doubt on the rush to deploy LLMs in the military and diplomatic domains, and calls for more research on their risks and limitations.
127

You are viewing a single comment

A glorified chatbot, in other words.

In other words, you don't really really know what LLMs are.

If one is feeling cynical; humans are chatbots in shoes.

Searle speaks frankly. Challenging those who deny the existence of consciousness, he wonders how to argue with them. "Should I pinch [those people] to remind them they are conscious?" remarks Searle. "Should I pinch myself and report the results in the Journal of Philosophy?"

One can only investigate their own consciousness, so we can't outrule chatbots are also having some subjective experience 🙃

I don't know if I love or hate your comment. (Yes, you're right, shut up.) Well played, Internet stranger.