Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/78357
Full metadata record
DC FieldValueLanguage
dc.date.accessioned2021-07-15T09:38:51Z-
dc.date.available2021-07-15T09:38:51Z-
dc.date.issued2016-
dc.identifier.citationTsukanova, A. (2016). A coarticulation model for articulatory speech synthesis (Master's dissertation).en_GB
dc.identifier.urihttps://www.um.edu.mt/library/oar/handle/123456789/78357-
dc.descriptionM.SC.ARTIFICIAL INTELLIGENCEen_GB
dc.description.abstractThe state-of-the-art techniques for speech synthesis rely either on concatenation of acoustic units taken from a vast pre-recorded speech database noting the relevant linguistic information or on statistical generation of the necessary acoustic parameters and using a speech production model. These approaches yield synthesis of good quality, but are purely technical solutions which bring no or very little information about the acoustics of speech or about how the articulators (mandible, tongue, lips, velum...) are controlled. In contrast, the articulatory approach generates the speech signal from the vocal tract shape and its modelled acoustic phenomena. The vocal tract deformation control comprises slow anticipation of the main constriction and fast and imperatively accurate aiming for consonants. The system predicts the sequence of vocal tract consecutive configurations from a sequence of phonemes of the French language to be articulated and a model of the coarticulation effects in it. We use static magnetic resonance imaging (MRI) captures of the vocal tract shape when producing phonemes in various contexts, thus following an approach by Birkholz (2013). The evaluation of the model is done both on the animated graphics representing the vocal tract shape evolution (how natural and efficient the movement is) and on the synthesised speech signals that are perceptively and-in terms of formants-qualitatively compared to identical utterances made Ly a human. Our results show that there are a lot of effects in the dynamic process of speech that manage to be reproduced by manipulating solely static data. We discuss generation of pure vowels, vowel-to-vowel and vowel-consonant-vowel transitions, and articulators' behaviour in phrases, report which acoustic properties have been rendered correctly and what could be the reasons for the system to fail to produce the desired result in other cases, and ponder how to reduce the after-effects of target-oriented moves to obtain a more gesture-like motion.en_GB
dc.language.isoenen_GB
dc.rightsinfo:eu-repo/semantics/restrictedAccessen_GB
dc.subjectFrench language -- Pronunciationen_GB
dc.subjectFrench language -- Phoneticsen_GB
dc.subjectSpeech processing systemsen_GB
dc.subjectMagnetic resonance imagingen_GB
dc.titleA coarticulation model for articulatory speech synthesisen_GB
dc.typemasterThesisen_GB
dc.rights.holderThe copyright of this work belongs to the author(s)/publisher. The rights of this work are as defined by the appropriate Copyright Legislation or as modified by any successive legislation. Users may access this work and can make use of the information contained in accordance with the Copyright Legislation provided that the author must be properly acknowledged. Further distribution or reproduction in any format is prohibited without the prior permission of the copyright holder.en_GB
dc.publisher.institutionUniversity of Maltaen_GB
dc.publisher.departmentFaculty of Information and Communication Technology. Department of Artificial Intelligenceen_GB
dc.description.reviewedN/Aen_GB
dc.contributor.creatorTsukanova, Anastasiia (2016)-
Appears in Collections:Dissertations - FacICT - 2016
Dissertations - FacICTAI - 2016

Files in This Item:
File Description SizeFormat 
M.SC.LANG.SCIENCE_TECH._Tsukanova_Anastasiia_2016.pdf
  Restricted Access
17.17 MBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.