[ExI] AI controlled weapons now being used in Ukraine
    BillK 
    pharos at gmail.com
       
    Fri Sep 12 15:47:47 UTC 2025
    
    
  
On Fri, 12 Sept 2025 at 11:54, BillK <pharos at gmail.com> wrote:
> Terminators: AI-driven robot war machines on the march
> Science fiction? Battle bots already used in Ukraine.
> Steven J. Vaughan-Nichols   Fri 12 Sep 2025
>
> <
> https://www.theregister.com/2025/09/12/terminators_aidriven_robot_war_machines/
> >
> ------------------------------
I asked CHATGPT-5 to explain more about the dangers of AI-controlled
weapons.
The answer was rather frightening.
BillK
CHATGPT-5:
Alright, let’s slow it down and walk through both sets of risks — the ones *The
Register* article explicitly pointed out, and the extra ones I added that
sit just under the surface.
------------------------------
*Risks the article itself highlighted*
   1.
   *Misidentification of targets*
   -
      AI systems rely on sensors and pattern recognition. In the chaos of
      battle (smoke, debris, unusual angles), they may confuse a
civilian car for
      a military one, or mistake surrendering soldiers for combatants.
      -
      Once that happens, the system can act faster than a human can
      intervene.
      2.
   *Delegation drift (“humans in theory, machines in practice”)*
   -
      Militaries insist on keeping humans in the loop, but in high-speed
      engagements, operators often defer to AI judgments because reaction times
      are too short.
      -
      This creates a *slippery slope*: formally, there’s oversight; in
      practice, the machine is calling the shot.
      3.
   *Encoding of human bias*
   -
      AI is trained or parameterized using human input. If biases are built
      in (for example, labeling certain groups as “suspicious”), the AI can
      reproduce or even amplify those prejudices.
      -
      Unlike a human officer, the AI can’t contextualize or show discretion
      once its rules are set.
      4.
   *Civilian harm*
   -
      Even small errors in target selection can have massive humanitarian
      impact. Drone swarms or autonomous munitions don’t “second-guess” — they
      act.
      -
      This risk is especially acute in urban battlefields where fighters
      and civilians are mixed.
      5.
   *Escalation risk*
   -
      The article mentions war-gaming studies where AI systems tend to
      escalate conflicts more aggressively than humans would.
      -
      Faster machine decision cycles can lead to tit-for-tat responses
      spiraling before humans even realize what’s happening.
------------------------------
*Additional risks that weren’t fully unpacked*
   1.
   *Error cascades*
   -
      One bad decision by an autonomous system can ripple outward. Example:
      a drone misidentifies a truck, destroys it, cutting off communication.
      Other systems interpret the blackout as enemy jamming, escalating further.
      -
      Machines chained together can amplify mistakes faster than humans can
      step in.
      2.
   *Adversarial misuse / hacking*
   -
      AI systems can be tricked (adversarial attacks) or hijacked. Imagine
      painting symbols on a roof that fool a drone’s vision into
thinking it’s a
      friendly target, or jamming GPS so it mis-navigates.
      -
      If hacked, the weapon can literally be turned against its owner.
      3.
   *Arms race dynamics*
   -
      Nations may feel pressured to *out-automate* rivals. The faster one
      side delegates control to machines, the more others will follow — not
      wanting to be left behind.
      -
      This speeds up the adoption of immature, poorly tested systems,
      raising accident risks.
      4.
   *Lower threshold for conflict*
   -
      Cheap, autonomous weapons (like drone swarms) reduce the political
      cost of war — fewer soldiers at risk. That makes it easier for leaders to
      say “yes” to escalation.
      -
      The psychological barrier to launching attacks gets thinner.
      5.
   *Accountability vacuum*
   -
      If an autonomous system commits a war crime (say, bombing civilians),
      who’s responsible?
      -
         The commander who deployed it?
         -
         The coder who wrote the targeting software?
         -
         The company that manufactured it?
         -
      Right now, international law has no clear answers, which risks a
      culture of impunity.
      6.
   *Norm erosion*
   -
      Even if current doctrine insists on human oversight, repeated use of
      semi-autonomous systems normalizes pushing the boundaries. Over time, the
      “line” of acceptable use moves quietly forward until full autonomy is the
      de facto norm.
------------------------------
*Why these matter together*
The picture isn’t just “AI might make mistakes.” It’s a *systemic risk
stack*:
   -
   AI speeds up decisions.
   -
   Humans struggle to stay in the loop.
   -
   Errors and biases scale up faster.
   -
   Opponents exploit weaknesses.
   -
   Political leaders face fewer costs to escalate.
   -
   Meanwhile, no one’s clearly accountable when it all goes wrong.
That’s the cocktail that makes AI-driven weapons scarier than just “smarter
missiles.
*------------------------------------------------------*
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20250912/b5cc0b9f/attachment.htm>
    
    
More information about the extropy-chat
mailing list