DFKI-LT - Towards Learning Human-Robot Dialogue Policies Combining Speech and Visual Beliefs

Heriberto Cuayahuitl, Ivana Kruijff-Korbayová
Towards Learning Human-Robot Dialogue Policies Combining Speech and Visual Beliefs
in: Ramón López-Cózar Delgado, Tetsunori Kobayashi (eds.):
2 Proceedings of IWSDS 2011: Workshop on Paralinguistic Information and its Integration in Spoken Dialogue Systems, Pages 133-140, Granada, Spain, Springer, 2011
 
We describe an approach for multi-modal dialogue strategy learning combining two sources of uncertainty: speech and gestures. Our approach represents the state-action space of a reinforcement learning dialogue agent with relational representations for fast learning, and extends it with belief state variables for dialogue control under uncertainty. Our approach is evaluated, using simulation, on a robotic spoken dialogue system for an imitation game of arm movements. Preliminary experimental results show that the joint optimization of speech and visual beliefs results in better overall system performance than treating them in isolation.
 
Files: BibTeX, hc-iwsds2011a.pdf