NewsWashington University Develops AntiFake: A Novel Tool to Combat Deepfake Speech Synthesis

Washington University Develops AntiFake: A Novel Tool to Combat Deepfake Speech Synthesis

Category articles

Recent advancements in generative artificial intelligence have led to significant progress in realistic speech synthesis. This technology, while beneficial for creating personalized voice assistants and enhancing communication tools, has also given rise to the threat of deepfakes, where synthesized speech can be misused for deceptive purposes. To address this issue, Ning Zhang, an assistant professor of computer science and engineering at the McKelvey School of Engineering, Washington University in St. Louis, has developed AntiFake, a novel defense mechanism to prevent unauthorized speech synthesis.

Introducing AntiFake at ACM Conference

Zhang presented AntiFake at the Association for Computing Machinery’s Conference on Computer and Communications Security in Copenhagen, Denmark, on November 27. This innovative tool represents a proactive approach to combating deepfake speech, different from traditional methods that primarily focus on post-attack detection and mitigation.

Proactive Defense Against Speech Synthesis

AntiFake utilizes adversarial AI techniques to make it challenging for AI tools to synthesize deceptive speech. It works by subtly distorting voice recordings in a way that remains imperceptible to human listeners but significantly impacts AI’s ability to generate synthetic audio. This approach ensures the security of voice data against potential misuse by cybercriminals.

High Effectiveness and Generalizability

To validate its effectiveness, AntiFake was tested against five state-of-the-art speech synthesizers and demonstrated a protection rate of over 95%, even against commercial synthesizers that it hadn’t previously encountered. This generalizability makes it a robust tool against a wide array of potential attackers and synthesis models.

User Accessibility Testing

The tool’s accessibility was also evaluated with 24 human participants, confirming that AntiFake can be used effectively by diverse populations. This aspect is crucial for ensuring widespread adoption and use of the tool.

Future Expansion and Applications

Currently, AntiFake is tailored to protect short clips of speech, targeting the most prevalent type of voice impersonation. However, Zhang envisions expanding its capabilities to protect longer recordings and potentially music. This expansion is part of the broader effort to combat disinformation and maintain the integrity of voice recordings.

Adapting to the Evolving AI Landscape

Zhang anticipates that while AI voice technology will continue to evolve, the strategy of using adversarial techniques against potential threats will remain effective. He acknowledges that AI systems are still vulnerable to adversarial perturbations, suggesting that adaptations in engineering specifics might be necessary to keep this defensive strategy successful in the future.

Michal Pukala
Electronics and Telecommunications engineer with Electro-energetics Master degree graduation. Lightning designer experienced engineer. Currently working in IT industry.

News