Pentagon officials have sounded the alarm about ‘unique classes of vulnerabilities for AI or autonomous systems,’ which they hope new research can fix. 

The program, dubbed Guaranteeing AI Robustness against Deception (GARD), has been tasked since 2022 with identifying how visual data or other electronic signals inputs for AI might be gamed by the calculated introduction of noise.   

Computer scientists with one of GARD’s defense contractors have experimented with kaleidoscopic patches designed to fool AI-based systems into making false IDs. 

‘You can essentially, by adding noise to an image or a sensor, perhaps break a downstream machine learning algorithm,’ as one senior Pentagon official managing the research explained Wednesday. 

The news comes as fears that the Pentagon has been ‘building killer robots in the basement’ have allegedly led to stricter AI rules for the US military — mandating that all systems must be approved before deployment.

Computer scientists with defense contractor MITRE Corp. managed to create visual noise that an AI mistook for apples on a grocery store shelf, a bag left behind outdoors, and even people

Computer scientists with defense contractor MITRE Corp. managed to create visual noise that an AI mistook for apples on a grocery store shelf, a bag left behind outdoors, and even people

Computer scientists with defense contractor MITRE Corp. managed to create visual noise that an AI mistook for apples on a grocery store shelf, a bag left behind outdoors, and even people 

A bus packed with civilians, for example, could be misidentified as a tank by an AI, if it were tagged with the right 'visual noise,' as one national security reporter with the site ClearanceJobs proposed as an example

A bus packed with civilians, for example, could be misidentified as a tank by an AI, if it were tagged with the right 'visual noise,' as one national security reporter with the site ClearanceJobs proposed as an example

Researchers with the modestly budgeted GARD program have spent $51,000 investigating visual and signal noise tactics since 2022, Pentagon audits show

Researchers with the modestly budgeted GARD program have spent $51,000 investigating visual and signal noise tactics since 2022, Pentagon audits show

A bus packed with civilians, for example, could be misidentified as a tank by an AI, if it were tagged with the right ‘visual noise,’ as one national security reporter with ClearanceJobs proposed as an example. The Pentagon program has spent $51,000 investigating since 2022

‘You can also with knowledge of that algorithm sometimes create physically realizable attacks,’ added that official, Matt Turek, deputy director for the Defense Advanced Research Projects Agency’s (DARPA’s) Information Innovation Office.

Technically, it is feasible to ‘trick’ an AI’s algorithm into mission critical errors — making the AI misidentify a variety of patterned patches or stickers for a real physical object that’s not really there.

A bus packed with civilians, for example, could be misidentified as a tank by an AI, if it were tagged with the right ‘visual noise,’ as one national security reporter with the site ClearanceJobs proposed as an example. 

Such cheap and lightweight ‘noise’ tactics, in short, could cause vital military AI to misclassify enemy combatants as allies, and vice versa, during a critical mission. 

Researchers with the modestly budgeted GARD program have spent $51,000 investigating visual and signal noise tactics since 2022, Pentagon audits show. 

A 2020 study by MITRE illustrated how visual noise, which can appear merely decorative or inconsequential to human eyes, like a 1990s 'Magic Eye' poster, can be interpreted as solid object by AI. Above, MITRE's visual noise tricks an AI into seeing apples

A 2020 study by MITRE illustrated how visual noise, which can appear merely decorative or inconsequential to human eyes, like a 1990s 'Magic Eye' poster, can be interpreted as solid object by AI. Above, MITRE's visual noise tricks an AI into seeing apples

A 2020 study by MITRE illustrated how visual noise, which can appear merely decorative or inconsequential to human eyes, like a 1990s ‘Magic Eye’ poster, can be interpreted as solid object by AI. Above, MITRE’s visual noise tricks an AI into seeing apples

US Deputy Assistant Secretary of Defense for Force Development and Emerging Capabilities Michael Horowitz explained at an event this January that the new Pentagon directive 'does not prohibit the development of any [AI] systems,' but will 'make clear what is and isn't allowed.' Above, a fictional killer robot from the Terminator film franchise

US Deputy Assistant Secretary of Defense for Force Development and Emerging Capabilities Michael Horowitz explained at an event this January that the new Pentagon directive 'does not prohibit the development of any [AI] systems,' but will 'make clear what is and isn't allowed.' Above, a fictional killer robot from the Terminator film franchise

US Deputy Assistant Secretary of Defense for Force Development and Emerging Capabilities Michael Horowitz explained at an event this January that the new Pentagon directive ‘does not prohibit the development of any [AI] systems,’ but will ‘make clear what is and isn’t allowed.’ Above, a fictional killer robot from the Terminator film franchise

What is public from their work includes a 2019 and 2020 study illustrating how visual noise, which can appear merely decorative or inconsequential to human eyes, like a 1990s ‘Magic Eye’ poster, can be interpreted as solid object by AI.

Computer scientists with defense contractor the MITRE Corporation managed to create visual noise that an AI mistook for apples on a grocery store shelf, a bag left behind outdoors, and even people.  

‘Whether that is physically realizable attacks or noise patterns that are added to AI systems,’ Turek said Wednesday, ‘the GARD program has built state-of-the-art defenses against those.’ 

‘Some of those tools and capabilities have been provided to CDAO [the Defense Department’s Chief Digital and AI Office],’ according to Turek.

The Pentagon formed the CDAO in 2022; it serves as a hub to facilitate faster adoption of AI and related machine-learning technologies across the military. 

The Department of Defense (DoD) recently updated its AI rules among ‘a lot of confusion about’ how it plans to use self-decision-making machines on the battlefield, according to US Deputy Assistant Secretary of Defense for Force Development and Emerging Capabilities Michael Horowitz 

Horowitz explained at an event this January that the ‘directive does not prohibit the development of any [AI] systems,’ but will ‘make clear what is and isn’t allowed’ and uphold a ‘commitment to responsible behavior,’ as it develops lethal autonomous systems.

While the Pentagon believes the changes should ease the public’s minds, some have said they are not ‘convinced’ by the efforts. 

Mark Brakel, director of the advocacy organization Future of Life Institute (FLI), told DailyMail.com this January: ‘These weapons carry a massive risk of unintended escalation.’

He explained that AI-powered weapons could misinterpret something, like a ray of sunlight, and perceive it as a threat, thus attacking foreign powers without cause, and without intentional adversarial ‘visual noise.’

Brakel said the result could be devastating because ‘without meaningful human control, AI-powered weapons are like the Norwegian rocket incident [a near nuclear armageddon] on steroids and they could increase the risk of accidents in hotspots such as the Taiwan Strait.’ 

Dailymail.com has reached out to the DoD for comment. 

The DoD has been aggressively pushing to modernize its arsenal with autonomous drones, tanks, and other weapons that select and attack a target without human intervention.

This post first appeared on Dailymail.co.uk

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Elon Musk Made an Unusual Request in Fallout With Top Lieutenant

Last year, Elon Musk parted ways with one of his top deputies…

Mind-bending optical illusion hides a SECRET image – but not everyone can see it

PREPARE for some eye trickery once more as a TikToker has revealed…

Mixed race people really are more attractive! Study confirms stars like Mariah Carey and Zayn Malik are rated as better looking

From Mariah Carey  to Zayn Malik, many of the most glamorous stars…

PS5 release latest – Playstation 5 on sale TODAY with Curry’s, Amazon, BT, John Lewis in stock & GAME, Tesco pre-orders

THE wait is over! The Playstation 5 went on sale in the…