Đang chuẩn bị nút TẢI XUỐNG, xin hãy chờ
Tải xuống
This paper presents a speech understanding component for enabling robust situated human-robot communication. The aim is to gain semantic interpretations of utterances that serve as a basis for multi-modal dialog management also in cases where the recognized word-stream is not grammatically correct. For the understanding process, we designed semantic processable units, which are adapted to the domain of situated communication. Our framework supports the specific characteristics of spontaneous speech used in combination with gestures in a real world scenario. . | Spontaneous Speech Understanding for Robust Multi-Modal Human-Robot Communication Sonja Hiiwel Britta Wrede Faculty of Technology Applied Computer Science Bielefeld University 33594 Bielefeld Germany shuewel bwrede@techfak.uni-bielefeld.de Abstract This paper presents a speech understanding component for enabling robust situated human-robot communication. The aim is to gain semantic interpretations of utterances that serve as a basis for multi-modal dialog management also in cases where the recognized word-stream is not grammatically correct. For the understanding process we designed semantic processable units which are adapted to the domain of situated communication. Our framework supports the specihc characteristics of spontaneous speech used in combination with gestures in a real world scenario. It also provides information about the dialog acts. Finally we present a processing mechanism using these concept structures to generate the most likely semantic interpretation of the utterances and to evaluate the interpretation with respect to semantic coherence. 1 Introduction Over the past years interest in mobile robot applications has increased. One aim is to allow for intuitive interaction with a personal robot which is based on the idea that people want to communicate in a natural way Breazeal et al. 2004 Daut-enhahn 2004 . Although often people use speech as the main modality they tend to revert to additional modalities such as gestures and mimics in face-to-face situations. Also they refer to objects This work has been supported by the European Union within the Cognitive Robot Companion COGNIRON project FP6-IST-002020 and by the German Research Foundation within the Graduate Program Task Oriented Communication . in the physical environment. Furthermore speech gestures and information of the environment are used in combination in instructions for the robot. When participants perceive a shared environment and act in it we call this communication situated Milde et