Banner image: Image of a polling place generated using artificial intelligence. (Credit: Adobe Stock)
On Aug. 18, former president and current presidential candidate Donald Trump posted an unusual endorsement to his social media account on Truth Social. Amid a series of photos, he included an image of pop megastar Taylor Swift wearing an Uncle Sam hat and declaring: “Taylor wants you to vote for Donald Trump.”
The problem: It wasn’t real. Swift hadn’t, and still hasn’t, endorsed a candidate for the 2024 presidential election. The image may have been generated by artificial intelligence.
Casey Fiesler, associate professor in the Department of Information Science at CU Boulder, sees the rise of AI in politics as a worrying trend. This month, for example, NPR reporter Huo Jingnan tested out Grok, a new AI platform launched by the social media company X. She was able to to create surprisingly realistic security camera images of people stuffing envelopes into ballot drop boxes in the dead of night.
“It’s not like fake images weren't a thing before this year,” Fiesler said. “The difference is that it's so much easier to do now. AI is democratizing this type of bad acting.”
To help voters navigate this new and perilous election information landscape, CU Boulder Today spoke to Fiesler and other experts in AI and media literacy. They include Kai Larsen, professor of information systems in the Leeds School of Business, and Toby Hopp, associate professor in the Department of Advertising, Public Relations and Media Design.
These experts discuss how you can find out if a photo you’re seeing online is the real deal—and how to talk to friends and family members who are spreading misinformation.
Yes, AI really is that good
In the past, AI-generated images often left behind “artifacts,” such as hands with six fingers, that eagle-eyed viewers could spot. But those sorts of mistakes are getting easier to fix in still images. Video is not far behind, said Fiesler, who covers the ethics of AI in a course she’s teaching this fall called “Ethical and Policy Dimensions of Information and Technology.”
“At some point soon, you will be able to see an AI-generated video of the head of the CDC giving a press conference, and it will totally fool you,” she said.
At the same time, the algorithms that govern social media platforms like TikTok and Instagram can trap users in downward spirals of misinformation, Larsen said. He’s the co-author of the 2021 book .
“Algorithms, at least historically, have been driving people into these rabbit holes,” Larsen said. “If you are willing to believe one piece of misinformation, then the algorithm is now finding out that you like conspiracy theories. So why not feed you more of them?”
Tech probably won’t save us
A range of companies now offer services that, they claim, can detect AI-generated content, including fake images. But just like human eyes, those tools can be easily tricked, Larsen said. Some critics of AI have also urged tech companies to add digital “watermarks” to AI content. These watermarks would flag photos or text that had originally come from an AI platform.
“The problem with watermarks is that they are often fairly easy to get rid of,” Larsen said. “Or you can just find another large language model that doesn’t use them.”Google it
When it comes to AI images, a little searching online can go a long way, Fiesler said.
Earlier in August, Trump’s campaign accused Kamala Harris’ team of using AI to make the crowd size look bigger in a photo of one of her rallies. Fiesler ran a quick Google image search on the image. She discovered that numerous news organizations had covered the same event, and dozens of other photos and videos existed, all showing the same large crowd.
“Google it,” Fiesler said. “Find out: Are news organizations writing about this same event? Do other photographs exist?”
Hopp, a scholar who studies fake news or what he prefers to call “countermedia,” cautions social media users to beware of posts that try to trigger our worst impulses. In 2016, troll farms in Russia posted thousands of misleading ads about the presidential election to social media. Many tried to tap into negative emotions, pitting Americans on the right and left against each other.
“We can evaluate a piece of information and ask ourselves: ‘Is this trying to make me angry? Is this trying to make me upset?’” Hopp said. “If so, we may want to ask: ‘Is there a possibility that this might be misleading?’”
What about friends and family?
It’s a familiar problem for many people—a friend or family member who won’t stop sharing misleading social media posts. Dealing with that kind of loved one can be a minefield, Hopp said. Research shows that simply challenging people on their false beliefs (say, that the Earth is flat) often won’t change their minds. It may even make them double down.
He and other researchers have experimented with giving social media users “” or basic info on how to tell fact from fiction. Such interventions can help, but not as much as Hopp would hope.
“I do think that sober, empathetic and caring discussions with those who are important to us about media literacy can be important for helping people use different strategies when they're on social media platforms,” he said. “But there’s no silver bullet.”
What can we do in the long-term?
Can anything help to slow the spread of misleading AI images online?
Fiesler sees an urgent need for the federal government to step in to regulate the rapidly growing AI industry. She said that a starting point could be the which the White House’s Office of Science and Technology Policy drafted in 2022.
This blueprint, which has not been passed into law, includes recommendations like: “You should know that an automated system is being used and understand how and why it contributes to outcomes that impact you.”
Hopp, for his part, believes that a lot of the responsibility for stopping political misinformation comes down to another group: politicians. It’s time to cool down the temperature of political discussions in the United States.
“There’s a role for our political leaders to discourage the use of hyper-partisan, divisive information,” Hopp said. “Embracing this type of misleading information creates conditions that are fairly ripe for its spread.”