Publication Details
Normalising Flows for Speaker and Language Recognition Backend
Prasad Amrutha (DCGM FIT BUT)
Motlíček Petr, doc. Ing., Ph.D. (DCGM FIT BUT)
Madikeri Srikanth (IDIAP)
Schuepbach Christof (armasuise)
Speaker recognition, Language Recognition
In this paper, we address the Gaussian distribution assumption made in PLDA, a popular back-end classifier used in Speaker and Language recognition tasks. We study normalizing flows, which allow using non-linear transformations and still obtain a model that can explicitly represent a probability density. The model makes no assumption about the distribution of the ob- servations. This alleviates the need for length normalization, a well known data preprocessing step used to boost PLDA performance. We demonstrate the effectiveness of this flow model on NIST SRE16, LRE17 and LRE22 datasets. We ob- serve that when applying length normalization, both the flow model and PLDA achieve similar EERs for SRE16 (11.5% vs 11.8%). However, when length normalization is not applied, the flow shows more robustness and offers better EERs (13.1% vs 17.1%). For LRE17 and LRE22, the best classification accu- racies (84.2%, 75.5%) are obtained by the flow model without any need for length normalization.
@INPROCEEDINGS{FITPUB13295, author = "Aleix Fontcuberta Espuna and Amrutha Prasad and Petr Motl\'{i}\v{c}ek and Srikanth Madikeri and Christof Schuepbach", title = "Normalising Flows for Speaker and Language Recognition Backend", pages = "74--80", booktitle = "Proceedings of Odyssey 2024: The Speaker and Language Recognition Workshop", year = 2024, location = "Quebec, CA", publisher = "International Speech Communication Association", doi = "10.21437/odyssey.2024-11", language = "english", url = "https://www.fit.vut.cz/research/publication/13295" }