Abstract
Unrehearsed spoken language often contains disfluencies. In order to correctly interpret a spoken utterance, any such disfluencies must be identified and removed or otherwise dealt with. Operating on transcripts of speech which contain disfluencies, we study the effect of language model and loss function on the performance of a linear reranker that rescores the 25-best output of a noisychannel model. We show that language models trained on large amounts of non-speech data improve performance more than a languagemodel trained on amoremodest amount of speech data, and that optimising f-score rather than log loss improves disfluency detection performance. Our approach uses a log-linear reranker, operating on the top n analyses of a noisy channel model. We use large language models, introduce new features into this reranker and examine different optimisation strategies. We obtain a disfluency detection f-scores of 0.838 which improves upon the current state-of-theart.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics, ACL-HLT 2011 |
| Place of Publication | Portland, OR |
| Publisher | Association for Computational Linguistics (ACL) |
| Pages | 703-711 |
| Number of pages | 9 |
| Volume | 1 |
| ISBN (Print) | 9781932432879 |
| Publication status | Published - 2011 |
| Event | 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, ACL-HLT 2011 - Portland, OR, United States Duration: 19 Jun 2011 → 24 Jun 2011 |
Other
| Other | 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, ACL-HLT 2011 |
|---|---|
| Country/Territory | United States |
| City | Portland, OR |
| Period | 19/06/11 → 24/06/11 |