The Canadian Security Intelligence Service (CSIS), Canada’s primary national intelligence agency, has expressed serious concerns about the escalating threat posed by disinformation campaigns utilizing artificial intelligence (AI) deepfakes. In a recent report, the CSIS highlighted the increasing realism of deepfakes, coupled with the challenge of recognizing or detecting them, as a potential danger to Canadians.
The report emphasized the potential harm caused by deepfakes, pointing to instances where these advanced AI technologies were used to target individuals. CSIS underscored the threat to democracy, citing the potential for certain actors to exploit uncertainty and propagate false information. The agency specifically mentioned the impact on governments, raising the concern that official content could be questioned and doubted if unable to prove its authenticity.
One notable example cited in the report was the deepfake videos featuring Elon Musk, which were used to deceive crypto investors. Since 2022, malicious actors have leveraged sophisticated deepfake videos to persuade unsuspecting individuals in the crypto community to part with their funds. CSIS emphasized the urgency of addressing these concerns, drawing attention to the potential consequences of governments not keeping pace with the evolving threat landscape.
Elon Musk's deep fake video promoting a new cryptocurrency scam going viral. The video claims that the trading platform is owned by Elon Musk, and offers 30% returns on crypto deposits. @elonmusk pic.twitter.com/iJeUvHYc5p— DogeDesigner (@cb_doge) May 24, 2022
The report also identified additional issues related to AI, including privacy violations, social manipulation, and bias. CSIS highlighted the need for swift and adaptive governmental policies, directives, and initiatives to effectively counter the growing realism of deepfakes and synthetic media.
To address these challenges, CSIS proposed collaborative efforts among partner governments, allies, and industry experts to tackle the global distribution of authentic information. The agency’s recommendation aligns with Canada’s recent commitment to addressing AI concerns on an international scale. On October 30, the Group of Seven (G7) industrial countries reached a consensus on an AI code of conduct designed to ensure the safe, secure, and trustworthy development of AI globally.
The 11-point code aims to balance the benefits of AI while mitigating the associated risks, reflecting the shared commitment of allied nations to navigate the challenges presented by advanced technologies.