‘The Gospel’: how Israel uses AI to select bombing targets in Gaza

flango@lemmy.eco.br to Technology@lemmy.world – 176 points –
‘The Gospel’: how Israel uses AI to select bombing targets in Gaza
theguardian.com
19

"The AI gave us bad advice. Our hands are clean."

There should be some sort of law where if you want to offload decisions to AI, the person who decides to let the AI make those decisions needs to step up to take full civil and criminal liability for everything it does.

so you only really need one martyr for the cause?

Yes, one person we can pin all of humanity's sins on, and then we just kill them. It's almost like a religious ritual.

No, every decision maker in the chain of command should be responsible. They should know what the intelligence is based on, if the people sharing the information are competent and should be validating the information.

Using AI to perform these tasks requires gross negligence at several stages. However, it does appear killing civilians and children is the intended outcome so negligence about AI is likely just a cover.

"as humans come to rely on these systems they become cogs in a mechanised process and lose the ability to consider the risk of civilian harm in a meaningful way."

that's not even an AI problem, that's a regular society problem

And since "ai" isn't actual artificial intelligence but rather a neural net trained on data from that society, it's only going to reinforce existing issues. Remember the articles about crime ai being racist? That's because they fed policing statistics as the training set, and uh.. I'm sure you know the rest.

This is fucking insane dystopian shit, it’s worse than I thought and has become real sooner than I thought it would, bloody hell.

Yeah we're getting really good at teaching computers to analyze satellite imagry and other forms of spatial data to find the spots we want. All we have to do is decide if we put green spaces, Walmarts or bombs in those spots.

This is the best summary I could come up with:


As Israel resumes its offensive after a seven-day ceasefire, there are mounting concerns about the IDF’s targeting approach in a war against Hamas that, according to the health ministry in Hamas-run Gaza, has so far killed more than 15,000 people in the territory.

The latest Israel-Hamas war has provided an unprecedented opportunity for the IDF to use such tools in a much wider theatre of operations and, in particular, to deploy an AI target-creation platform called “the Gospel”, which has significantly accelerated a lethal production line of targets that officials have compared to a “factory”.

The Guardian can reveal new details about the Gospel and its central role in Israel’s war in Gaza, using interviews with intelligence sources and little-noticed statements made by the IDF and retired officials.

This article also draws on testimonies published by the Israeli-Palestinian publication +972 Magazine and the Hebrew-language outlet Local Call, which have interviewed several current and former sources in Israel’s intelligence community who have knowledge of the Gospel platform.

In the IDF’s brief statement about its target division, a senior official said the unit “produces precise attacks on infrastructure associated with Hamas while inflicting great damage to the enemy and minimal harm to non-combatants”.

Multiple sources told the Guardian and +972/Local Call that when a strike was authorised on the private homes of individuals identified as Hamas or Islamic Jihad operatives, target researchers knew in advance the number of civilians expected to be killed.


The original article contains 1,734 words, the summary contains 241 words. Saved 86%. I'm a bot and I'm open source!

Can AI target what story might next outrage people on the political fringes? I'd like to not hear about Al Jazeera and the Daily Caller any more.

It may shock you to hear that centrists and capitalists don't have your best interests at heart.

Oh my goodness unplug from the matrix. Live free kind internet soul.