Abstract
Natural language generators are faced with a multitude of different decisions during their generation process. We address the joint optimisation of navigation strategies and referring expressions in a situated setting with respect to task success and human-likeness. To this end, we present a novel, comprehensive framework that combines supervised learning, Hierarchical Reinforcement Learning and a hierarchical Information State. A human evaluation shows that our learnt instructions are rated similar to human instructions, and significantly better than the supervised learning baseline.
Original language | English |
---|---|
Title of host publication | Proceedings of the SIGDIAL 2011 Conference |
Subtitle of host publication | The 12th Annual Meeting of the Special Interest Group on Discourse and Dialogue |
Place of Publication | Stroudsburg, PA |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 78-87 |
Number of pages | 10 |
ISBN (Print) | 9781937284107 |
Publication status | Published - 2011 |
Event | 12th Annual Meeting of the Special Interest Group on Discourse and Dialogue, SIGDIAL 2011 - Portland, OR, United States Duration: 17 Jun 2011 → 18 Jun 2011 |
Other
Other | 12th Annual Meeting of the Special Interest Group on Discourse and Dialogue, SIGDIAL 2011 |
---|---|
Country/Territory | United States |
City | Portland, OR |
Period | 17/06/11 → 18/06/11 |