Abstract
A method for automatic audiovisual transcription of speech employing: acoustic and visual speech representations is developed. It adopts a combining of audio and visual modalities, which provide a synergy effect in terms of speech recognition accuracy. To establish a robust solution, basic research concerning the relation between the allophonic variation of speech, i.e. the changes in the articulatory setting of speech organs for the same phoneme produced in different phonetic environments and the objective signal parameters (both audio and video) is carried out. The method is sensitive to minute allophonic detail as well as to accentual differences. It is shown that by using the analysis of video signals together with the acoustic signal, speech transcription can be performed more accurately and robustly than by using the acoustic modality alone. In particular, various features extracted from the visual signal are tested for their abilities to encode allophonic variations in pronunciation. New methods for modeling the accentual and allophonic variation of speech are developed.
Citations
-
0
CrossRef
-
0
Web of Science
-
0
Scopus
Authors (3)
Cite as
Full text
full text is not available in portal
Keywords
Details
- Category:
- Conference activity
- Type:
- publikacja w wydawnictwie zbiorowym recenzowanym (także w materiałach konferencyjnych)
- Published in:
-
Journal of the Acoustical Society of America
no. 26,
pages 1 - 15,
ISSN: 0001-4966 - Title of issue:
- Proceedings of Meetings on Acoustics: 171st Meeting of the Acoustical Society of America strony 1 - 15
- Language:
- English
- Publication year:
- 2016
- Bibliographic description:
- Ciszewski T., Czyżewski A., Kostek B.: Methodology and technology for the polymodal allophonic speech transcription// Proceedings of Meetings on Acoustics: 171st Meeting of the Acoustical Society of America/ : , 2016, s.1-15
- DOI:
- Digital Object Identifier (open in new tab) 10.1121/2.0000300
- Verified by:
- Gdańsk University of Technology
seen 120 times