Publication Details
Comprehensive Multiparametric Analysis of Human Deepfake Speech Recognition
Firc Anton, Ing. (DITS FIT BUT)
Šalko Milan, Ing. (DITS FIT BUT)
Prudký Daniel, Bc. (FIT BUT)
Radačovská Karolína, Bc. (FIT BUT)
Hanáček Petr, doc. Dr. Ing. (DITS FIT BUT)
Deepfake, Synthetic speech, Deepfake detection, Human perception, Speech quality, Cybersecurity
In this paper, we undertake a novel two-pronged investigation into the human recognition of deepfake speech, addressing critical gaps in existing research. First, we pioneer an evaluation of the impact of prior information on deepfake recognition, setting our work apart by simulating real-world attack scenarios where individuals are not informed in advance of deepfake exposure. This approach simulates the unpredictability of real-world deepfake attacks, providing unprecedented insights into human vulnerability under realistic conditions. Second, we introduce a novel metric to evaluate the quality of deepfake audio. This metric facilitates a deeper exploration into how the quality of deepfake speech influences human detection accuracy. By examining both the effect of prior knowledge about deepfakes and the role of deepfake speech quality, our research reveals the importance of these factors, contributes to understanding human vulnerability to deepfakes, and suggests measures to enhance human detection skills.
@ARTICLE{FITPUB13163, author = "Kamil Malinka and Anton Firc and Milan \v{S}alko and Daniel Prudk\'{y} and Karol\'{i}na Rada\v{c}ovsk\'{a} and Petr Han\'{a}\v{c}ek", title = "Comprehensive Multiparametric Analysis of Human Deepfake Speech Recognition", pages = "1--25", journal = "Eurasip Journal on Image and Video Processing", volume = 2024, number = 1, year = 2024, ISSN = "1687-5281", doi = "10.1186/s13640-024-00641-4", language = "english", url = "https://www.fit.vut.cz/research/publication/13163" }