Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/92466
Title: Data augmentation for speech recognition in Maltese : a low-resource perspective
Authors: Mena, Carlos
DeMarco, Andrea
Borg, Claudia
van der Plas, Lonneke
Gatt, Albert
Keywords: Maltese language
Speech synthesis
Linguistics
Automatic speech recognition
Natural language processing (Computer science)
Computational linguistics
Speech processing systems
Issue Date: 2021-11-15
Publisher: arXiv
Citation: Mena, C., DeMarco, A., Borg, C., van der Plas, L., & Gatt, A. (2021). Data augmentation for speech recognition in Maltese : a low-resource perspective. arXiv preprint arXiv:2111.07793.
Abstract: Developing speech technologies is a challenge for low-resource languages for which both annotated and raw speech data is sparse. Maltese is one such language. Recent years have seen an increased interest in the computational processing of Maltese, including speech technologies, but resources for the latter remain sparse. In this paper, we consider data augmentation techniques for improving speech recognition for such languages, focusing on Maltese as a test case. We consider three different types of data augmentation: unsupervised training, multilingual training and the use of synthesized speech as training data. The goal is to determine which of these techniques, or combination of them, is the most effective to improve speech recognition for languages where the starting point is a small corpus of approximately 7 hours of transcribed speech. Our results show that combining the three data augmentation techniques studied here lead us to an absolute WER improvement of 15% without the use of a language model.
URI: https://www.um.edu.mt/library/oar/handle/123456789/92466
Appears in Collections:Scholarly Works - InsSSA

Files in This Item:
File Description SizeFormat 
Data_Augmentation_for_Speech_Recognition_in_Maltese_A_Low_Resource_Perspective(2021).pdf321.52 kBAdobe PDFView/Open


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.