A multimodal real-time MRI articulatory corpus for speech research

Shrikanth Narayanan*, Erik Bresch, Prasanta Ghosh, Louis Goldstein, Athanasios Katsamanis, Yoon Kim, Adam Lammert, Michael Proctor, Vikram Ramanarayanan, Yinghua Zhu

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference proceeding contributionpeer-review

60 Citations (Scopus)

Abstract

We present MRI-TIMIT: a large-scale database of synchronized audio and real-time magnetic resonance imaging (rtMRI) data for speech research. The database currently consists of speech data acquired from two male and two female speakers of American English. Subjects' upper airways were imaged in the midsagittal plane while reading the same 460 sentence corpus used in the MOCHA-TIMIT corpus [1]. Accompanying acoustic recordings were phonemically transcribed using forced alignment. Vocal tract tissue boundaries were automatically identified in each video frame, allowing for dynamic quantification of each speaker's midsagittal articulation. The database and companion toolset provide a unique resource with which to examine articulatory-acoustic relationships in speech production.

Original languageEnglish
Title of host publicationProceedings of the Annual Conference of the International Speech Communication Association
Place of PublicationBaixas
PublisherInternational Speech Communication Association (ISCA)
Pages837-840
Number of pages4
Publication statusPublished - 2011
Externally publishedYes
Event12th Annual Conference of the International-Speech-Communication-Association 2011 (INTERSPEECH 2011) - Florence, Italy
Duration: 27 Aug 201131 Aug 2011

Conference

Conference12th Annual Conference of the International-Speech-Communication-Association 2011 (INTERSPEECH 2011)
CountryItaly
CityFlorence
Period27/08/1131/08/11

Fingerprint Dive into the research topics of 'A multimodal real-time MRI articulatory corpus for speech research'. Together they form a unique fingerprint.

Cite this