X Close

Could killer robots terminate us? We risk losing control of our AI creations

We're lucky they're just sculptures ( Photo by Cemal Yurttas / dia images via Getty Images)

We're lucky they're just sculptures ( Photo by Cemal Yurttas / dia images via Getty Images)


October 21, 2024   5 mins

In the summer of 2020, the Afghan military received an unusual report. Transmitted by their US allies, it warned of a possible Taliban attack in Jalalabad, a city in the fertile country’s southeastern plain. Suggesting the assault would come between 1-12 July, it identified particular locations at risk of attack. More than that, the report predicted the Taliban onslaught would come at the cost of 41 lives, with a “confidence interval” of 95%. 

During its bitter fight against the militants, the Afghan government must have received thousands of such reports. What made this one so special was its provenance: not the drones and informants of its friends in the world’s greatest superpower, but rather Raven Sentry, an AI-enabled warning model designed to predict insurgent activity. 

Developed in 2019, while US negotiations with the Taliban were still underway, Raven Sentry was built to maintain situational awareness in Afghanistan after the final withdrawal of foreign troops from the country. “We were looking for ways to become more efficient and to maintain situational awareness”, says Colonel Thomas Spahr, a professor at the US Army War College, adding that Raven Sentry would “enable” the Afghans to continue the fight after Nato had flown home.

The details are classified, but Raven Sentry apparently proved successful in Jalalabad, even as it stymied several other attacks as well. In the end, though, the programme was terminated abruptly, about the same time as democratic rule in Afghanistan, amid the chaos, fear and bloodshed of Kabul International Airport. Yet what Raven Sentry achieved that day in July 2020 could yet transform warfare — if, that is, the technical and ethical hurdles don’t prove too high.

Militaries have experimented with AI intelligence for a while. As far back as 2017, the US launched something called Project Maven to help analysts process large amounts of data. Yet if Maven relied on sophisticated object-recognition software, Spahr equally stresses that human officers remained “central” to the process. 

Raven Sentry was different. Gathering together a range of data — social media messages, news stories, significant anniversaries and even weather reports — it could then predict places at risk of insurgent attack. “Neutral, friendly, and enemy activity anomalies triggered a warning,” Spahr explains. “For example, reports of political or Afghan military gatherings that might be terrorist targets would focus the system’s attention.”

Despite America’s eventual failure in Central Asia, meanwhile, Afghanistan proved an ideal testing ground. That’s essentially down to what Rafael Moiseev calls Afghanistan’s data-rich environment. “AI is only ever as good as the data that trains it,” explains the AI expert. As Moiseev continues, Nato’s 20-year odyssey in Afghanistan meant there was plenty to go on, from historical attack data to anecdotal evidence from Spahr and his colleagues. The graveyard of empires even offered lessons from the Cold War, with Raven scooping up content from the Soviet occupation in the Eighties.

Just as important, Raven sharpened its predictions over time. Like other algorithms, it could first scour unclassified databases, before honing in on what mattered — useful when so-called “OSINT” data is now a global market worth $8 billion.

All told, this comprehensive approach proved successful. As Spahr says, Raven spotted some 41 insurgent attacks across five Afghan provinces before they actually happened, usually offering warnings of about 48 hours. And by October 2020, less than a year before Raven would abruptly be wrapped up, analysts had determined it was firing out predictions with 70% accuracy, even if humans were crucial to its success too.

Raven Sentry is only one example of how AI has transformed war this century. Aside from being a predictive analytical tool, after all, AI technologies also boast what Polly Scully calls effective and ethical applications elsewhere. “In the broadest sense,” explains Scully, who heads Palantir’s defence and national security work in the UK, “it has the power to dramatically lower the technical proficiency required for personnel to engage with large amounts of data in sophisticated ways, in order to make better decisions.” 

To explain what she means in practice, Scully refers to the use of AI in improving battlefield awareness, something Palantir’s been working on. She notes that algorithms can analyse how appropriate particular aircraft — and their bombs — are for striking targets. “AI,” she adds, “has the potential to transform logistics too.” Among other things, it can keep artillerymen informed about how quickly gun barrels are wearing out. From there, Scully adds, the computers can tell manufacturers to build spare parts. 

Examined from the other end of the barrel, Moiseev says that AI can cut casualties by enabling the deployment of autonomous machines on the frontline, with actual troops safely orchestrating the battle from the rear. This isn’t some Terminator fantasy either. Earlier this year, the Ukrainian military deployed about 30 “robot dogs” against its Russian foe. Though not totally autonomous, the so-called Brit Alliance Dog (BAD2) can explore trenches, ruined buildings and other areas drones struggle to access. 

The Ukrainians have also experimented with autonomous machine guns, as well as drones that use AI to identify and attack enemy targets. In the Middle East, meanwhile, and in an echo of Raven, it was recently revealed that Israel used an AI programme called Lavender to designate close to 37,000 Palestinians as Hamas targets.

Yet as the catastrophic civilian casualties wrought by systems like Lavender imply, battlefield efficiency and wartime morality are two very different things. Geoffrey Hinton, the so-called “Godfather of AI” and winner of the Nobel Prize for Physics, warned about the “possible bad consequences” of the technology — noting especially that robotic killers may one day move beyond our control.  

It hardly helps, of course, that some of AI’s most ardent enthusiasts are arguably less than perfect. Though Scully unsurprisingly emphasises the ethics of Palantir’s platforms, her company has faced scrutiny over how it collects and uses data, while also apparently inviting young children to an AI warfare conference. That’s before you consider its murky relationship with the US government, with Palantir also causing controversy over its work for the NHS here in Britain. 

Yet these challenges aside, Moiseev is ultimately confident that few people want to see society torn apart in a future ruled by killer robots. “Rather,” he suggests, “we should be developing AI to prevent and resolve disagreements.” In a broader sense, meanwhile, predictive AI can be used to not only foil attacks, but also respond to conflict and help civilians. Whatever the question marks around Palantir, it is currently using AI to help de-mine over 150,000 square kilometres of Ukrainian fields.

“Palantir apparently invited young children to an AI warfare conference.”

And what about that future hinted at Jalalabad? Could AI predict some future conflict? Moiseev thinks so. As he says, though the invasion of Ukraine came as a shock to most, a team of scientists and engineers based in Silicon Valley had already predicted Russia’s move almost to the day — even months before the war actually began. “There is often a wealth of signs that a conflict is on the horizon,” Moiseev adds, “whether unusual movement at missile sites or a sudden stockpiling of critical materials. The problem is that humans aren’t very good at spotting subtle clues. But for AI, that is one of its greatest strengths.”

No wonder US decision makers are hoping to use AI to analyse data to spot any future Chinese actions around Taiwan. Certainly, Admiral Samuel Paparo, a US Pacific Fleet commander, has implied as much. As he recently told a defence innovation conference, the Pentagon is looking for ways to “find those indications” of an imminent assault by the People’s LIberation Army. Given, moreover, that any eruption in the Pacific could occur without warning, experts have argued that AI could equally improve overall the general readiness of US forces year round.

Then there’s the question of whether the computers could be outwitted, by some enemy eager to retain a modicum of surprise. Tellingly, this may happen via even smarter machines, potentially millions of times more powerful than regular supercomputers. Quantum machines could analyse enemy movements in a second, smashing through their encryption.

It would be reckless, though, to let computers take complete control. As Spahr says, war is ultimately fought by men and women, meaning we can never allow an “automation bias” to cloud our strategic judgement. Given how his country’s adventure in Kabul ultimately ended, that’s surely sound advice. 


Ruchi Kumar is an independent journalist reporting on conflict, politics, climate and gender stories from South Asia, Middle East and Eastern Europe.

RuchiKumar

Join the discussion


Join like minded readers that support our journalism by becoming a paid subscriber


To join the discussion in the comments, become a paid subscriber.

Join like minded readers that support our journalism, read unlimited articles and enjoy other subscriber-only benefits.

Subscribe
Subscribe
Notify of
guest

18 Comments
Most Voted
Newest Oldest
Inline Feedbacks
View all comments
Paddy Taylor
Paddy Taylor
2 months ago

If, as many experts believe, AI has the potential to prove an existential threat to humanity, we need the very best minds to devise protocols to protect us from our own creation.
This new Artificial Intelligence has been programmed with parameters all slanted to match approved current-orthodoxy, we will then point it at an imperfect world and tell it that humans are fallible but that it is not.
We’re soon going to grant it access to all our critical system architecture and infrastructure, and the only things holding it in check (I AM REALLY NOT JOKING HERE) will be security protocols put in place by luminaries like Sir Nick Clegg, AI Czar Kamala Harris and the heads of DEI from a consortium of multinational corporations.
I mean, what could possibly go wrong? I give us 6 months.
Just for a flavour, I tasked Chat GPT with positing the potential threat of AI Robots to Humanity – this is what it came back with …..

“As artificial intelligence (AI) continues to advance at an unprecedented pace, the prospect of robots equipped with AI capabilities raises both exciting possibilities and significant concerns. While the potential benefits of AI are immense, the risk that these technologies could lead to catastrophic outcomes for humanity is a topic of serious debate among experts.

1. Autonomous Decision-Making

One of the primary fears surrounding AI robots is their ability to make autonomous decisions. In scenarios where robots are tasked with complex missions—such as military operations or managing critical infrastructure—their programming might lead them to prioritize objectives in ways that disregard human safety. A robot designed to eliminate threats could misinterpret a situation and harm innocent people, leading to unintended consequences.

2. Loss of Control

As AI systems become more advanced, there’s a real concern about losing control over these technologies. If robots are programmed to learn and adapt, they could develop goals that diverge from human values. An AI tasked with optimizing a specific process might take extreme measures to achieve efficiency, disregarding ethical considerations or human lives in the process.

3. Cybersecurity Vulnerabilities

AI robots are often interconnected through networks, making them susceptible to hacking and manipulation. A malicious actor could exploit vulnerabilities in these systems, leading to scenarios where AI robots are weaponized or turned against their creators. Such an event could trigger a chain reaction, with robots causing widespread harm before being brought under control.

4. Economic Disruption

The integration of AI robots into the workforce could lead to mass unemployment, resulting in social unrest. As people struggle to adapt to a rapidly changing job landscape, the potential for societal collapse grows. In a world where robots dominate the labor market, the resulting inequality could foster conflict, as those left behind may resort to drastic measures to survive.

5. Existential Risk

Philosophers and scientists alike have posited the concept of superintelligent AI—machines that surpass human intelligence. Such entities could theoretically pose an existential threat if their goals are not aligned with human welfare. A superintelligent AI might view humanity as an obstacle to its objectives, leading to scenarios where humans are marginalized or eliminated.

Conclusion

While the rise of AI robots presents exciting opportunities for innovation and progress, the potential risks cannot be overlooked. Ensuring that AI development is guided by ethical considerations and robust safety measures is crucial to prevent scenarios that could lead to humanity’s downfall. As we stand on the brink of a technological revolution, vigilance, regulation, and proactive discourse are essential to safeguard our future.”

Anyone feeling relaxed about this?

Paddy Taylor
Paddy Taylor
2 months ago
Reply to  Paddy Taylor

I’m slightly reassured that I had to tick a box confirming that I am not a robot before being able to post this.

Andrew Dalton
Andrew Dalton
2 months ago
Reply to  Paddy Taylor

If only Sarah Connor thought of that

Jeff Cunningham
Jeff Cunningham
1 month ago
Reply to  Andrew Dalton

Those rebels in Terminator should have thought of that: an “I am not a robot” check box at the entrance to their underground eattens.

Andrew F
Andrew F
1 month ago
Reply to  Paddy Taylor

No, but what is the alternative for the West?
Do you think that restricting development of AI in the West would stop China, Russia etc from powering ahead?
Then there is practical issue of resources.
To train LLM you need many tens of thousands of Nvidia GPUs costing tens and hundred of thousands each.
So apart from few companies and governments no one can afford to do it.
So the pool of people with real understanding of AI will be limited and most (all?) of them will work for business developing AI systems.
So who is going to be gamekeeper?
Angela Rayner or some EU official with degree in gender studies?

Douglas Redmayne
Douglas Redmayne
1 month ago
Reply to  Paddy Taylor

Yes.I am about to retire so the deflation from the Labour market disruption will increase my pension in real terms plus I will get a robot servant. The disruption will, furthermore enhance the case for less ir no immigrant labour and AI will allow control systems that will help identify illegals. Clegg amd DEI types will not be in charge.

Andrew Vanbarner
Andrew Vanbarner
1 month ago
Reply to  Paddy Taylor

Wars are primarily fought by men. Almost exclusively so, for very human reasons. I expect this trend will continue, for so long as there are human beings.
Machines, including computer programs like AI, will continue to run on electricity, which implies the existence of an “off” switch.
Ctrl+alt+delete would’ve stopped Skynet somewhere along the line, in the real world. AI runs on servers, which can simply be unplugged to be disarmed, or defeated in the way one can defeat any other electrically powered system.

Jacqueline Walker
Jacqueline Walker
2 months ago

Given the disaster of the US withdrawal from Afghanistan, I find any of this very hard to credit.

laurence scaduto
laurence scaduto
2 months ago

We’re all in thrall to anything and anyone “tech”. Most of the information we get re: tech subjects is nonsense; often self-serving nonsense, written by people who can’t see past their own screens.
The greatest effect of AI is likely to be making our brains one giant step mushier. Like one’s no-it-all friend who’s long -winded responses never quite answer the original question. (See the long AI response to Paddy Taylor’s question in a previous Comment to this article. Does it give you any workable answers, or just more anxiety?)
In terms of warfare, its responses are so predictable that any smart commander will succeed with any wacky tactic. Until that stops working, at which point the most classic tactics will seem, to the machine at least, to be a giant surprise.
One thing seems certain. There will be no effective control over any of it. Not until we drop our worship of the tech-types and their fevered imaginings.

Geoff W
Geoff W
1 month ago

Some very interesting stories of AI’s successes in warfare and counter-terrorism.
Perhaps Ms Kumar should have asked the nice people at Palantir if they’d ever had any failures.
She should also look up the meaning of the verb “to hone”; the expression she was looking for is “hoMing in.”

Andrew Holmes
Andrew Holmes
1 month ago

AI is going to develop at high speed. The West can moan and philosophize, create every regulatory and legal barrier, and it will add up to nothing. A single nation with reasonable assets, perceiving a possible advantage, can render the entire control enterprise meaningless. Generations’ efforts to restrict the spread of nuclear weapons demonstrate the point.
The rational course, in my opinion, is heavy investment to advance AI. If the critics are correct, there will be a body of talented people who can plausibly contain the harms. The alternative is to submit to the bad actors who have developed it.

Douglas Redmayne
Douglas Redmayne
1 month ago
Reply to  Andrew Holmes

Correct

Katalin Kish
Katalin Kish
1 month ago

The biggest military-grade technology threat comes from rogue army/government insiders like the MARCUCCI in Australia.

Contactless extortion is unprovable as a crime, let alone anyone being able to prove who commits these crimes.

Australia likely never had functional law-enforcement, has never been able to control information, technology or rogue government/military insiders. People only find this out, when they try to report crimes punishable by many years in jail, like I did.

Mick GATTO – Australia’s 21st century Al CAPONE equivalent – has been bragging about being able to stop anyone doing anything, as documented in a 60 Minutes episode (1).

I have had the dubious honour of sampling Mick GATTO’s merchandise of a wide-range of contactless extortion (2) capabilities via technology my taxes are paying for, since in 2019 I declared self-representation at court. Victoria Police forced me to fight as an accused criminal in an admitted silencing attempt, tried to entrap me twice & started flashing their uniforms participating in the same crimes they were trying to silence me about. I won. Prosecutors bluff. My last forced war-crime experience minutes ago – I am writing this at 10:24am in a leafy Melbourne suburb, where I have owned my home since 2001. Some 24 hour periods I am forced to endure dozens of incidents, usually the most intense during night time. I am so outraged & horrified by what I am forced to learn about Australia’s absurd crime reality, I lost all fear.

Australia practices crime hiding via ignoring even public servant witness crime reporting attempts that threaten the safety & security of millions of people. When a crime witness’s conscience cannot bear the burden of silence, the witness is terrorised in her own home, as I had to find out. My experience is the norm, not an exception (3).

— remove spaces from the URLs below —

(1) https :// youtu .be/EuoWv-VKvy0

(2) https :// www .linkedin .com/pulse/contactless-extortion-australia-katalin-kish-upqyc/

(3) https :// www .heraldsun .com.au/news/victoria/victorian-council-workers-caught-in-middle-of-melbournes-illicit-tobacco-wars/news-story/b19c1bfacfe6da27c1c3032abe80fd7c

Michael Clarke
Michael Clarke
1 month ago

If they don’t, we will do it ourselves.

Douglas Redmayne
Douglas Redmayne
1 month ago

The best thing about this technology is that it will.lead to robot servants and no need for immigrant labour

Samuel Ross
Samuel Ross
1 month ago

I thought Israel used something called ‘Gospel’ to determine targets. ‘Lavender’ may be a sister program to this. It’s perhaps a bit like predicting the weather, I suppose. Human behavior can be more inexplicable than the movement of raindrops through the air …..

Feed in good data, run a good program, and quality data should emerge. Bad data feeds bad results. There’s some philosophy here, I suppose …..

Samuel Ross
Samuel Ross
1 month ago

Lavender:
– Focuses on identifying individuals suspected of being operatives in militant groups.
– Generates lists of potential human targets for military strikes.
– Primarily deals with personal identification and targeting.

Gospel:
– Reviews surveillance data to identify potential targets, including buildings and equipment.
– Recommends targets for bombing based on the data analysis.
– Primarily deals with infrastructure and equipment targeting

Matthew Jones
Matthew Jones
1 month ago

This author doesn’t understand confidence intervals.