Publication Details
Text Augmentation for Language Models in High Error Recognition Scenario
data augmentation, error simulation, language modeling, automatic speech recognition
In this paper, we explore several data augmentation strategies for training of language models for speech recognition. We compare augmentation based on global error statistics with one based on unigram statistics of ASR errors and with labelsmoothing and its sampled variant. Additionally, we investigate the stability and the predictive power of perplexity estimated on augmented data. Despite being trivial, augmentation driven by global substitution, deletion and insertion rates achieves the best rescoring results. On the other hand, even though the associated perplexity measure is stable, it gives no better prediction of the final error rate than the vanilla one. Our best augmentation scheme increases the WER improvement from second-pass rescoring from 1.1% to 1.9% absolute on the CHiMe-6 challenge.
@INPROCEEDINGS{FITPUB12606, author = "Karel Bene\v{s} and Luk\'{a}\v{s} Burget", title = "Text Augmentation for Language Models in High Error Recognition Scenario", pages = "1872--1876", booktitle = "Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH", journal = "Proceedings of Interspeech - on-line", volume = 2021, number = 8, year = 2021, location = "Brno, CZ", publisher = "International Speech Communication Association", ISSN = "1990-9772", doi = "10.21437/Interspeech.2021-627", language = "english", url = "https://www.fit.vut.cz/research/publication/12606" }