News

Biden robocall: Audio deepfake fuels election disinformation fears

By AFP

February 5, 2024 10:05 AM


Representational image

Twitter Share Facebook Share WhatsApp Share

The 2024 White House race faces the prospect of a firehose of AI-enabled disinformation, with a robocall impersonating US President Joe Biden already stoking particular alarm about audio deepfakes.

"What a bunch of malarkey," said the phone message, digitally spoofing Biden's voice and echoing one of his signature phrases.

The robocall urged New Hampshire residents not to cast ballots in the Democratic primary last month, prompting state authorities to launch a probe into possible voter suppression.

It also triggered demands from campaigners for stricter guardrails around generative artificial intelligence tools or an outright ban on robocalls.

Disinformation researchers fear rampant misuse of AI-powered applications in a pivotal election year thanks to proliferating voice cloning tools, which are cheap and easy to use and hard to trace.

"This is certainly the tip of the iceberg," Vijay Balasubramaniyan, chief executive and co-founder of cybersecurity firm Pindrop, told AFP.

"We can expect to see many more deepfakes throughout this election cycle."

A detailed analysis published by Pindrop said a text-to-speech system developed by the AI voice cloning startup ElevenLabs was used to create the Biden robocall.

The scandal comes as campaigners on both sides of the US political aisle harness advanced AI tools for effective campaign messaging and as tech investors pump millions of dollars into voice cloning startups.

Balasubramaniyan refused to say whether Pindrop had shared its findings with ElevenLabs, which last month announced a financing round from investors that, according to Bloomberg News, gave the firm a valuation of $1.1 billion.

ElevenLabs did not respond to repeated AFP requests for comment. Its website leads users to a free text-to-speech generator to "create natural AI voices instantly in any language."

Under its safety guidelines, the firm said users were allowed to generate voice clones of political figures such as Donald Trump without their permission if they "express humor or mockery" in a way that makes it "clear to the listener that what they are hearing is a parody, and not authentic content."

- 'Electoral chaos' -

US regulators have been considering making AI-generated robocalls illegal, with the fake Biden call giving the effort new impetus.

"The political deepfake moment is here," said Robert Weissman, president of the advocacy group Public Citizen.

"Policymakers must rush to put in place protections or we're facing electoral chaos. The New Hampshire deepfake is a reminder of the many ways that deepfakes can sow confusion."

Researchers fret the impact of AI tools that create videos and text so seemingly real that voters could struggle to decipher truth from fiction, undermining trust in the electoral process.

But audio deepfakes used to impersonate or smear celebrities and politicians around the world have sparked the most concern.

"Of all the surfaces -- video, image, audio -- that AI can be used for voter suppression, audio is the biggest vulnerability," Tim Harper, a senior policy analyst at the Center for Democracy & Technology, told AFP.

"It is easy to clone a voice using AI, and it is difficult to identify."

- 'Election integrity' -

The ease of creating and disseminating fake audio content complicates an already hyperpolarized political landscape, undermining confidence in the media and enabling anyone to claim that fact-based "evidence has been fabricated," Wasim Khaled, chief executive of Blackbird.AI, told AFP.

Such concerns are rife as the proliferation of AI audio tools outpaces detection software.

China's ByteDance, owner of the wildly popular platform TikTok, recently unveiled StreamVoice, an AI tool for real-time conversion of a user's voice to any desired alternative.

"Even though the attackers used ElevenLabs this time, it is likely to be a different generative AI system in future attacks," Balasubramaniyan said.

"It is imperative that there are enough safeguards available in these tools."

Balasubramaniyan and other researchers recommended building audio watermarks or digital signatures into tools as possible protections as well as regulation that makes them available only for verified users.

"Even with those actions, detecting when these tools are used to generate harmful content that violates your terms of service is really hard and really expensive," Harper said.

"(It) requires investment in trust and safety and a commitment to building with election integrity centred as a risk."


AFP


Most Read

  1. Solar panel rate sinks massively in Lahore Solar panel rate sinks massively in Lahore
  2. Hina Rizvi ties the knot with Ammar Ahmed Khan Hina Rizvi ties the knot with Ammar Ahmed Khan
  3. Police claim girl seen in sub-inspector’s video was his second wife Police claim girl seen in sub-inspector’s video was his second wife
  4. Foreigners safe after suicide-bomber targets their van in Karachi Foreigners safe after suicide-bomber targets their van in Karachi
  5. LHC orders steel mill demolition over pollution LHC orders steel mill demolition over pollution
  6. Mehar Bano and Mamya Shajaffar’s bold dance videos invite public's wrath Mehar Bano and Mamya Shajaffar’s bold dance videos invite public's wrath

Opinion

  1. Insights into the Pakistan Stock Exchange's Recent Record High Triumph
    Insights into the Pakistan Stock Exchange's Recent Record High Triumph

    By Zulfiqar Ali Mir

  2. IMEC to sabotage CPEC
    IMEC to sabotage CPEC

    By Dr Asif Channer

  3. 1947 TO FORM 47
    1947 TO FORM 47

    By Dr Asif Channer

  4. Beijing wants to further highlight industrial sector in its country and take scientific innovation to new heights....
    Beijing wants to further highlight industrial sector in its country and take scientific innovation to new heights....

    By Ali Ramay

  5. Global race: China will reduce its unnecessary expenses
    Global race: China will reduce its unnecessary expenses

    By Ali Ramay

  6. Channer Pir: The Great Saint of Cholistan
    Channer Pir: The Great Saint of Cholistan

    By Dr Asif Channer