Satellite images showing the expansion of large detention camps in Xinjiang, China, between 2016 and 2018 provided some of the strongest evidence of a government crackdown on more than a million Muslims, triggering international condemnation and sanctions.

Other aerial images—of nuclear installations in Iran and missile sites in North Korea, for example—have had a similar impact on world events. Now, image-manipulation tools made possible by artificial intelligence may make it harder to accept such images at face value.

In a paper published online last month, University of Washington professor Bo Zhao employed AI techniques similar to those used to create so-called deepfakes to alter satellite images of several cities. Zhao and colleagues swapped features between images of Seattle and Beijing to show buildings where there are none in Seattle and to remove structures and replace them with greenery in Beijing.

Zhao used an algorithm called CycleGAN to manipulate satellite photos. The algorithm, developed by researchers at UC Berkeley, has been widely used for all sorts of image trickery. It trains an artificial neural network to recognize the key characteristics of certain images, such as a style of painting or the features on a particular type of map. Another algorithm then helps refine the performance of the first by trying to detect when an image has been manipulated.

A map (upper left) and satellite image (upper right) of Tacoma. The lower images have been altered to make Tacoma look more like Seattle (lower left) and Beijing (lower right). 

Courtesy of Zhao et al., 2021, Journal of Cartography and Geographic Information Science

As with deepfake video clips that purport to show people in compromising situations, such imagery could mislead governments or spread on social media, sowing misinformation or doubt about real visual information.

“I absolutely think this is a big problem that may not impact the average citizen tomorrow but will play a much larger role behind the scenes in the next decade,” says Grant McKenzie, an assistant professor of spatial data science at McGill University in Canada, who was not involved with the work.

“Imagine a world where a state government, or other actor, can realistically manipulate images to show either nothing there or a different layout,” McKenzie says. “I am not entirely sure what can be done to stop it at this point.”

A few crudely manipulated satellite images have already spread virally on social media, including a photograph purporting to show India lit up during the Hindu festival of Diwali that was apparently touched up by hand. It may be just a matter of time before far more sophisticated “deepfake” satellite images are used to, for instance, hide weapons installations or wrongly justify military action.

Gabrielle Lim, a researcher at Harvard Kennedy School’s Shorenstein Center who focuses on media manipulation, says maps can be used to mislead without AI. She points to images circulated online suggesting that Alexandria Ocasio-Cortez was not where she claimed to be during the Capitol insurrection on January 6, as well as Chinese passports showing a disputed region of the South China Sea as part of China. “No fancy technology, but it can achieve similar objectives,” Lim says.

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Apple warns all iPhone owners to never make ’sell’ mistake – it’s too dangerous

IF you ever plan on selling your iPhone, Apple has some serious…

A US Propaganda Operation Hit Russia and China With Memes

This week, former Twitter chief security officer Peiter “Mudge” Zatko filed an…

Brexit: EE, O2, Three and Vodafone WON’T charge for mobile data in Europe after transition period 

Mobile providers EE, O2, Three and Vodafone have all confirmed that they…

Google plots cheap new rival to Amazon Fire TV Stick – and fans are hoping it’ll solve ‘lousy’ storage issue

GOOGLE is reportedly planning a brand new budget TV streaming device to…