Violent extremists who lack the means to carry out an attack in Canada could compensate by perpetrating hoaxes with the help of artificial intelligence, says a newly released analysis.
The May report by the federal Integrated Terrorism Assessment Centre, obtained through the Access to Information Act, warns that such visual trickery, known as a deepfake, poses 鈥渁 persistent threat to public safety.鈥
The assessment centre鈥檚 report was prompted by an image of dark smoke rising near the U.S. Pentagon that appeared May 22 on social media, causing stocks to drop temporarily. Officials confirmed there was no emergency.
Synthetic images, video and audio are becoming easier to generate through applications driven by artificial intelligence, allowing people to spread false information and sow confusion.
The centre, which employs members of the security and intelligence community, predicted threat actors would 鈥渁lmost certainly鈥 create deepfake images depicting Canadian interests in the coming year, given the available tools and prevalence of misinformation and disinformation.
Rapid Response Mechanism Canada, a federal unit that tracks foreign information manipulation and interference, recently highlighted such an episode, saying it likely originated with the Chinese government.
The foreign operation, which began in August, employed a network of new or hijacked social media accounts to post comments and videos that called into question the political and ethical standards of various MPs, using a popular Chinese-speaking figure in Canada, RRM Canada said.
The terrorism assessment centre analysis says extremists could use deepfakes to advocate for violence, promote specific narratives, cause panic, tarnish reputations, and erode trust in government and societal institutions.
鈥淗oaxes provide violent extremists with an effective technique to disrupt daily life or to intimidate targeted groups or individuals, including by potentially diverting security resources from their regular duties,鈥 the report says.
鈥淐ontinued deepfake hoaxes could result in a competitive environment among threat actors, where the goal is to cause increasing real-world impacts, such as economic harm.鈥
Violent extremists with 鈥渓imited capabilities鈥 are more likely to use hoaxes than actors who are capable of conducting 鈥渕ore direct actions,鈥 the report concludes.
鈥淒eepfake media will make it more difficult for authorities to respond to, prioritize and investigate violent extremist threats.鈥
In May, the Canadian Security Intelligence Service invited experts and security practitioners to a workshop to explore the threats posed by such disinformation technologies.
A resulting report, based on papers presented at the event, said terrorist organizations 鈥渟urely recognize鈥 the potential of employing deepfakes in the spread of propaganda and co-ordination of attacks.
鈥淭he looming spectre of deepfakes presents an unprecedented threat to national security. The rapid evolution and proliferation of this technology demands nothing short of a resolute and comprehensive response,鈥 the report said.
Democracies must invest in cutting-edge deepfake detection technologies that can unmask digital imposters, as well as criminalize the creation and dissemination of deepfakes, it added.
However, the battle against deepfakes cannot be won through technology and legislation alone, the report cautioned.
鈥淐itizens must be armed with the power of critical thinking and media literacy, thereby empowering them to discern truth from fabrication. By fostering a society that is professionally skeptical, informed and resilient, governments can build a shield against the corrosive effects of deepfakes.鈥
Still, that task might still prove challenging.
Just this month, the Canadian Centre for Cyber Security noted in a report on threats to democratic institutions that 鈥渢he capacity to generate deepfakes exceeds our ability to detect them.鈥
鈥淐urrent publicly available detection models struggle to reliably distinguish between deepfakes and real content.鈥
READ ALSO:
READ ALSO: