[ad_1]
Just days before Slovakia’s national election last fall, a mysterious voice recording began spreading a lie online.
The manipulated file made it sound like Michal Simecka, leader of the Progressive Slovakia party, was discussing buying votes with a local journalist. But the conversation never happened; the file was later debunked as a “deepfake” hoax.
On election day, Simecka lost to the pro-Kremlin populist candidate Robert Fico in a tight race.
While it’s nearly impossible to determine whether the deepfake file contributed to the final results, the incident points to growing fears about the effect products of artificial intelligence are having on democracy around the world — and in Canada.
“This is what we fear … that there could be a foreign interference so grave that then the electoral roll results are brought into question,” Caroline Xavier, head of the Communications Security Establishment (CSE) — Canada’s cyber intelligence agency — told CBC News.
“We know that misinformation and disinformation is already a threat to democratic processes. This will potentially add to that amplification. That is quite concerning.”
Those concerns are playing out around the world this year in what’s being described as democracy’s biggest test in decades.
Billions of people in more than 40 countries are voting in elections this year — including what’s expected to be a bitterly disputed U.S. presidential contest. Canadians could be headed to the polls this year or next, depending on how much longer the Liberal government’s deal with the NDP holds up.
“I don’t think anybody is really ready,” said Hany Farid, a professor at the University of California-Berkeley specializing in digital forensics.
Farid said he sees two main threats emerging from the collision of generative AI content and politics. The first is its effect on politicians — on their ability to deny reality.
“If your prime minister or your president or your candidate gets caught saying something actually offensive or illegal … you don’t have to cop to anything anymore,” he said.
“That’s worrisome to me, where nobody has to be held accountable for anything they say or do anymore, because there’s the spectre of deepfakes hanging over us.”
The second threat, he said, is already playing out: the spread of fake content designed to harm individual candidates.
“If you’re trying to create a 10 second hot mic of the prime minister saying something inappropriate, that’ll take me two minutes to do. And very little money and very little effort and very little skill,” Farid said.
“It doesn’t matter if you correct the record 12 hours later. The damage has been done. The difference between the candidates is typically in the tens of thousands of votes. You don’t have to move millions of votes.”
Cyber intelligence agency prepares for ‘the worst’
The consequences are very much on the minds of the experts working within the glass walls of CSE’s 72,000-square metre headquarters in Ottawa.
Last month, the foreign signals intelligence agency released a public report warning that bad actors will use AI tools to manipulate voters.
“Canada is not immune. We know this could happen,” said Xavier. “We anticipate the worst. I’m hoping it won’t happen, but we’re ready.
“There’s lots of work we continue to need to do with regards to education, and … citizenship literacy. Absolutely, I think we’re ready. Because this is what we trained for, this is what we get ready for, this is why we develop our people.”
CSE’s preparations for an AI assault on Canada’s elections include the authority to knock misleading content offline.
“Could we potentially use defensive cyber operations should the need arise? Absolutely,” Xavier said. “Our minister had authorized them leading up to the 2019 and the 2021 election. We did not have to use it. But in anticipation of the upcoming election, we would do the same. We’d be ready.”
Xavier said Canada’s continued use of paper ballots in national elections affords it a degree of protection from online interference.
CSE, the Canadian Security Intelligence Service (CSIS), the RCMP and Global Affairs Canada will also feed intelligence about attempts to manipulate voters to decision makers in the federal government before and during the next federal election campaign.
The federal government established the Critical Election Incident Public Protocol in 2019 to monitor and alert the public to credible threats to Canada’s elections. The team is a panel of top public servants tasked with determining whether incidents of interference meet the threshold for warning the public.
The process has been criticized by opposition MPs and national security experts for not flagging fake content and foreign interference in the past two elections. Last year, a report reviewing the panel’s work suggested the government should consider amending the threshold so that the panel can issue an alert when there is evidence of a “potential impact” on an election.
The Critical Election Incident Public Protocol likely will be studied by the public inquiry probing election interference later this month.
CSE warns that AI technology is advancing at a pace that ensures it won’t be able to detect every single deceptive video or image deployed to exploit voters — and some people inevitably will fall for fake AI-generated content before they head to the ballot box.
According to its December report, CSE believes that it is “very likely that the capacity to generate deepfakes exceeds our ability to detect them” and that “it is likely that influence campaigns using generative AI that target voters will increasingly go undetected by the general public.”
Xavier said training the public to spot counterfeit online content must be part of efforts to ensure Canada is ready for its next federal campaign.
“The reality of it is … yes, it would be great to say that there’s this one tool that’s going to help us decipher the deepfake. We’re not there yet,” she said. “And I don’t know that that’s the focus we should have. Our focus should truly be in creating professional scepticism.
“I’m hopeful that the social media platforms will also play a role and continue to educate people with regards to what they should be looking at, because that’s where we know a lot of our young people hang out.”
A spokesperson for YouTube said that since November 2023, it’s been requiring content creators to disclose any altered or synthetic content. Meta, which owns Facebook and Instagram, said this year that advertisers also will have to disclose in certain cases their use of AI or other digital techniques to create or alter advertising on political or social issues.
Parliament isn’t moving fast enough, MP says
It’s not enough to put Conservative MP Michelle Rempel Garner at ease.
“I have over a decade worth of speeches that are on the internet … It’d be very easy for somebody to put together a deepfake video of me,” she said.
She said she wants to see a stronger response to the threat from the federal government.
“I mean, we haven’t even dealt with telephone scams as a country, right? We really haven’t dealt with beta-version phone scams. And now here we are with very sophisticated technology that anybody can access and come up with very realistic videos that are indistinguishable [from] the real thing,” said the MP for Calgary Nose Hill.
Those fears convinced Rempel Garner to help set up a bipartisan parliamentary caucus on emerging technology to educate MPs from all parties about the dangers, and opportunities, of artificial intelligence.
“There’s some really tough questions that we’re going to have to ask ourselves about how we deal with this, but also protect free speech. It’s just something that really makes my skin crawl. And I just feel the sense of urgency, that we’re not moving forward with it fast enough,” she said.
U.S. President Joe Biden, meanwhile, has introduced a new set of government-drafted standards on watermarking AI-generated content to help users distinguish between real and phoney content.
Rempel Garner said a watermark initiative is something Canada also could do “in short order.”
A spokesperson for Public Safety Minister Dominic LeBlanc suggested the government will have more to say on this subject at some point.
“We are concerned about the role that artificial intelligence could play in helping persons or entities knowingly spread false information that could disrupt the conduct of a federal election, or undermine its legitimacy,” said Jean-Sébastien Comeau.
“We are working on measures to address this issue and will have more to say in due course.”
AI companies need to take responsibility, expert says
Farid said regulations and legislation alone will not tame the “big bad internet out there.”
Companies that permit users to create fake content could also require that such content include a durable watermark identifying it as AI-generated, he said.
“I would like to see the open AI companies be more responsible in terms of how they are developing and deploying their technologies. But I’m also realistic about the way capitalism in the world works,” Farid said.
Farid also called for making date-time-place watermarks standard on phones.
“The idea is that if I pick up my phone here and I take a video of police violence, or human rights violations or a candidate saying something inappropriate, this device can record and authenticate where I am, when I was there, who I am and what I recorded,” he said.
Farid said he sees a way forward through a combination of technological solutions, regulatory pressure, public education and after-the-fact analysis of questionable content.
“I think all of those solutions start to bring some trust back into our online world, but all of them need to be pushed on simultaneously,” he said.
Friends don’t let friends fall for deepfakes
Scott DeJong is focused on the public education part of that equation. The PhD candidate at Montreal’s Concordia University created a board game to show how disinformation and conspiracy theories spread and has taught young people and foreign militaries how to play.
As AI technology advances, it might soon be impossible to teach people not to fall for fake content during elections. But DeJong said you can still teach people to recognize content as misleading.
“If you see a headline, and the headline is really emotional, or it’s manipulative, those are good signs [that], well, this content is probably at least misleading,” he said.
“My actual advice for people during … election times is to try to watch things live. Because it’s a lot harder to try to see the deepfakes or the false content when you’re watching the live version,” he said.
He also said Canadians can do their part by reaching out to friends and families when they post disinformation — especially when those loved ones refuse to engage with reputable mainstream news sources.
“The optimist in me likes to think that no one is too far gone,” he said.
“Don’t go in there accusing them or blaming them, but [ask] them questions as to why they put that content up. Just keep asking, why did you think that post was important? What about that post did you find interesting? What in that content engaged you?
“From there, you can peel back layers of the ideas and perspectives that led to them sharing that.”
[ad_2]
Source link