The misuse of generative AI (genAI) has raised significant ethical and trust issues. To mitigate this, substantial focus has been placed on detecting generated media, including fake audio. In this paper, we examine the efficacy of state-of-the-art fake audio detection methods under real-world conditions. By analyzing typical audio alterations of transmission pipelines, we identify several vulnerabilities: (1) minimal changes such as sound level variations can bias detection performance, (2) inevitable physical effects such as background noise lead to classifier failures, (3) classifiers struggle to generalize across different datasets, and (4) network degradation affects the overall detection performance. Our results indicate that existing detectors have major issues in differentiating between real and fake audio in practical applications and that significant improvements are still necessary for reliable detection in real-world environments.
History
Primary Research Area
Threat Detection and Defenses
Name of Conference
International Conference on Machine Learning (ICML)
Journal
ICML 2024 Next Generation of AI Safety Workshop
BibTeX
@conference{Shaw:Nassi:Schönherr:2024,
title = "Generated Audio Detectors are Not Robust in Real-World Conditions",
author = "Shaw, Soumya" AND "Nassi, Ben" AND "Schönherr, Lea",
year = 2024,
month = 6,
journal = "ICML 2024 Next Generation of AI Safety Workshop"
}