Home Business Deepfake Maps Could Really Mess With Your Sense of the World

Deepfake Maps Could Really Mess With Your Sense of the World

0
Deepfake Maps Could Really Mess With Your Sense of the World

[ad_1]

Satellite images showing the expansion of large detention camps in Xinjiang, China, between 2016 and 2018 provided some of the strongest evidence of a government crackdown on more than a million Muslims, triggering international condemnation and sanctions.

Other aerial images—of nuclear installations in Iran and missile sites in North Korea, for example—have had a similar impact on world events. Now, image-manipulation tools made possible by artificial intelligence may make it harder to accept such images at face value.

In a paper published online last month, University of Washington professor Bo Zhao employed AI techniques similar to those used to create so-called deepfakes to alter satellite images of several cities. Zhao and colleagues swapped features between images of Seattle and Beijing to show buildings where there are none in Seattle and to remove structures and replace them with greenery in Beijing.

Zhao used an algorithm called CycleGAN to manipulate satellite photos. The algorithm, developed by researchers at UC Berkeley, has been widely used for all sorts of image trickery. It trains an artificial neural network to recognize the key characteristics of certain images, such as a style of painting or the features on a particular type of map. Another algorithm then helps refine the performance of the first by trying to detect when an image has been manipulated.

A map (upper left) and satellite image (upper right) of Tacoma. The lower images have been altered to make Tacoma look more like Seattle (lower left) and Beijing (lower right). 

Courtesy of Zhao et al., 2021, Journal of Cartography and Geographic Information Science

As with deepfake video clips that purport to show people in compromising situations, such imagery could mislead governments or spread on social media, sowing misinformation or doubt about real visual information.

“I absolutely think this is a big problem that may not impact the average citizen tomorrow but will play a much larger role behind the scenes in the next decade,” says Grant McKenzie, an assistant professor of spatial data science at McGill University in Canada, who was not involved with the work.

“Imagine a world where a state government, or other actor, can realistically manipulate images to show either nothing there or a different layout,” McKenzie says. “I am not entirely sure what can be done to stop it at this point.”

A few crudely manipulated satellite images have already spread virally on social media, including a photograph purporting to show India lit up during the Hindu festival of Diwali that was apparently touched up by hand. It may be just a matter of time before far more sophisticated “deepfake” satellite images are used to, for instance, hide weapons installations or wrongly justify military action.

Gabrielle Lim, a researcher at Harvard Kennedy School’s Shorenstein Center who focuses on media manipulation, says maps can be used to mislead without AI. She points to images circulated online suggesting that Alexandria Ocasio-Cortez was not where she claimed to be during the Capitol insurrection on January 6, as well as Chinese passports showing a disputed region of the South China Sea as part of China. “No fancy technology, but it can achieve similar objectives,” Lim says.



[ad_2]

Source link

LEAVE A REPLY

Please enter your comment!
Please enter your name here