Abstract
The relationship between how people describe objects and when they choose to point is complex and likely to be influenced by factors related to both perceptual and discourse context. In this paper, we explore these interactions using machine-learning on a dialogue corpus, to identify multimodal referential strategies that can be used in automatic multimodal generation. We show that the decision to use a pointing gesture depends on features of the accompanying description (especially whether it contains spatial information), and on visual properties, especially distance or separation of a referent from its previous referent.
Original language | English |
---|---|
Title of host publication | Proceedings of the 25th International Conference on Computational Linguistics (COLING '14) |
Number of pages | 10 |
Place of Publication | Dublin, Ireland |
Publisher | Association for Computational Linguistics |
Publication date | 2014 |
Pages | 2007-2017 |
Publication status | Published - 2014 |
Event | Coling 2014 - Dublin, Ireland Duration: 23 Aug 2014 → 29 Aug 2014 |
Conference
Conference | Coling 2014 |
---|---|
Country/Territory | Ireland |
City | Dublin |
Period | 23/08/2014 → 29/08/2014 |