In the summer of 2020, the Afghan military received an unusual report. Transmitted by their US allies, it warned of a possible Taliban attack in Jalalabad, a city in the fertile country’s southeastern plain. Suggesting the assault would come between 1-12 July, it identified particular locations at risk of attack. More than that, the report predicted the Taliban onslaught would come at the cost of 41 lives, with a “confidence interval” of 95%.
During its bitter fight against the militants, the Afghan government must have received thousands of such reports. What made this one so special was its provenance: not the drones and informants of its friends in the world’s greatest superpower, but rather Raven Sentry, an AI-enabled warning model designed to predict insurgent activity.
Developed in 2019, while US negotiations with the Taliban were still underway, Raven Sentry was built to maintain situational awareness in Afghanistan after the final withdrawal of foreign troops from the country. “We were looking for ways to become more efficient and to maintain situational awareness”, says Colonel Thomas Spahr, a professor at the US Army War College, adding that Raven Sentry would “enable” the Afghans to continue the fight after Nato had flown home.
The details are classified, but Raven Sentry apparently proved successful in Jalalabad, even as it stymied several other attacks as well. In the end, though, the programme was terminated abruptly, about the same time as democratic rule in Afghanistan, amid the chaos, fear and bloodshed of Kabul International Airport. Yet what Raven Sentry achieved that day in July 2020 could yet transform warfare — if, that is, the technical and ethical hurdles don’t prove too high.
Militaries have experimented with AI intelligence for a while. As far back as 2017, the US launched something called Project Maven to help analysts process large amounts of data. Yet if Maven relied on sophisticated object-recognition software, Spahr equally stresses that human officers remained “central” to the process.
Raven Sentry was different. Gathering together a range of data — social media messages, news stories, significant anniversaries and even weather reports — it could then predict places at risk of insurgent attack. “Neutral, friendly, and enemy activity anomalies triggered a warning,” Spahr explains. “For example, reports of political or Afghan military gatherings that might be terrorist targets would focus the system’s attention.”
Join the discussion
Join like minded readers that support our journalism by becoming a paid subscriber
To join the discussion in the comments, become a paid subscriber.
Join like minded readers that support our journalism, read unlimited articles and enjoy other subscriber-only benefits.
SubscribeIf, as many experts believe, AI has the potential to prove an existential threat to humanity, we need the very best minds to devise protocols to protect us from our own creation.
This new Artificial Intelligence has been programmed with parameters all slanted to match approved current-orthodoxy, we will then point it at an imperfect world and tell it that humans are fallible but that it is not.
We’re soon going to grant it access to all our critical system architecture and infrastructure, and the only things holding it in check (I AM REALLY NOT JOKING HERE) will be security protocols put in place by luminaries like Sir Nick Clegg, AI Czar Kamala Harris and the heads of DEI from a consortium of multinational corporations.
I mean, what could possibly go wrong? I give us 6 months.
Just for a flavour, I tasked Chat GPT with positing the potential threat of AI Robots to Humanity – this is what it came back with …..
Anyone feeling relaxed about this?
I’m slightly reassured that I had to tick a box confirming that I am not a robot before being able to post this.
If only Sarah Connor thought of that
Those rebels in Terminator should have thought of that: an “I am not a robot” check box at the entrance to their underground eattens.
No, but what is the alternative for the West?
Do you think that restricting development of AI in the West would stop China, Russia etc from powering ahead?
Then there is practical issue of resources.
To train LLM you need many tens of thousands of Nvidia GPUs costing tens and hundred of thousands each.
So apart from few companies and governments no one can afford to do it.
So the pool of people with real understanding of AI will be limited and most (all?) of them will work for business developing AI systems.
So who is going to be gamekeeper?
Angela Rayner or some EU official with degree in gender studies?
Yes.I am about to retire so the deflation from the Labour market disruption will increase my pension in real terms plus I will get a robot servant. The disruption will, furthermore enhance the case for less ir no immigrant labour and AI will allow control systems that will help identify illegals. Clegg amd DEI types will not be in charge.
Wars are primarily fought by men. Almost exclusively so, for very human reasons. I expect this trend will continue, for so long as there are human beings.
Machines, including computer programs like AI, will continue to run on electricity, which implies the existence of an “off” switch.
Ctrl+alt+delete would’ve stopped Skynet somewhere along the line, in the real world. AI runs on servers, which can simply be unplugged to be disarmed, or defeated in the way one can defeat any other electrically powered system.
Given the disaster of the US withdrawal from Afghanistan, I find any of this very hard to credit.
We’re all in thrall to anything and anyone “tech”. Most of the information we get re: tech subjects is nonsense; often self-serving nonsense, written by people who can’t see past their own screens.
The greatest effect of AI is likely to be making our brains one giant step mushier. Like one’s no-it-all friend who’s long -winded responses never quite answer the original question. (See the long AI response to Paddy Taylor’s question in a previous Comment to this article. Does it give you any workable answers, or just more anxiety?)
In terms of warfare, its responses are so predictable that any smart commander will succeed with any wacky tactic. Until that stops working, at which point the most classic tactics will seem, to the machine at least, to be a giant surprise.
One thing seems certain. There will be no effective control over any of it. Not until we drop our worship of the tech-types and their fevered imaginings.
Some very interesting stories of AI’s successes in warfare and counter-terrorism.
Perhaps Ms Kumar should have asked the nice people at Palantir if they’d ever had any failures.
She should also look up the meaning of the verb “to hone”; the expression she was looking for is “hoMing in.”
AI is going to develop at high speed. The West can moan and philosophize, create every regulatory and legal barrier, and it will add up to nothing. A single nation with reasonable assets, perceiving a possible advantage, can render the entire control enterprise meaningless. Generations’ efforts to restrict the spread of nuclear weapons demonstrate the point.
The rational course, in my opinion, is heavy investment to advance AI. If the critics are correct, there will be a body of talented people who can plausibly contain the harms. The alternative is to submit to the bad actors who have developed it.
Correct
The biggest military-grade technology threat comes from rogue army/government insiders like the MARCUCCI in Australia.
Contactless extortion is unprovable as a crime, let alone anyone being able to prove who commits these crimes.
Australia likely never had functional law-enforcement, has never been able to control information, technology or rogue government/military insiders. People only find this out, when they try to report crimes punishable by many years in jail, like I did.
Mick GATTO – Australia’s 21st century Al CAPONE equivalent – has been bragging about being able to stop anyone doing anything, as documented in a 60 Minutes episode (1).
I have had the dubious honour of sampling Mick GATTO’s merchandise of a wide-range of contactless extortion (2) capabilities via technology my taxes are paying for, since in 2019 I declared self-representation at court. Victoria Police forced me to fight as an accused criminal in an admitted silencing attempt, tried to entrap me twice & started flashing their uniforms participating in the same crimes they were trying to silence me about. I won. Prosecutors bluff. My last forced war-crime experience minutes ago – I am writing this at 10:24am in a leafy Melbourne suburb, where I have owned my home since 2001. Some 24 hour periods I am forced to endure dozens of incidents, usually the most intense during night time. I am so outraged & horrified by what I am forced to learn about Australia’s absurd crime reality, I lost all fear.
Australia practices crime hiding via ignoring even public servant witness crime reporting attempts that threaten the safety & security of millions of people. When a crime witness’s conscience cannot bear the burden of silence, the witness is terrorised in her own home, as I had to find out. My experience is the norm, not an exception (3).
— remove spaces from the URLs below —
(1) https :// youtu .be/EuoWv-VKvy0
(2) https :// www .linkedin .com/pulse/contactless-extortion-australia-katalin-kish-upqyc/
(3) https :// www .heraldsun .com.au/news/victoria/victorian-council-workers-caught-in-middle-of-melbournes-illicit-tobacco-wars/news-story/b19c1bfacfe6da27c1c3032abe80fd7c
If they don’t, we will do it ourselves.
The best thing about this technology is that it will.lead to robot servants and no need for immigrant labour
I thought Israel used something called ‘Gospel’ to determine targets. ‘Lavender’ may be a sister program to this. It’s perhaps a bit like predicting the weather, I suppose. Human behavior can be more inexplicable than the movement of raindrops through the air …..
Feed in good data, run a good program, and quality data should emerge. Bad data feeds bad results. There’s some philosophy here, I suppose …..
Lavender:
– Focuses on identifying individuals suspected of being operatives in militant groups.
– Generates lists of potential human targets for military strikes.
– Primarily deals with personal identification and targeting.
Gospel:
– Reviews surveillance data to identify potential targets, including buildings and equipment.
– Recommends targets for bombing based on the data analysis.
– Primarily deals with infrastructure and equipment targeting
This author doesn’t understand confidence intervals.