/ プログラム/ 発表一覧/ 著者一覧/ 企業展示一覧/ jsai2010ホーム /

1J1-OS13-4 Detection of Robot-Directed Speech by Situated Understanding in Physical Interaction

06月09日(Wed) 14:35〜18:30 J会場(2F-練習室3)
1J1-OS13 オーガナイズドセッション「OS-13 マルチモーダル・セマンティック・インタラクション」

演題番号1J1-OS13-4
題目Detection of Robot-Directed Speech by Situated Understanding in Physical Interaction
著者左 祥(国際電気通信基礎技術研究所,京都工芸繊維大学 情報工学)
岩橋 直人(国際電気通信基礎技術研究所,(独)情報通信研究機構 )
田口 亮((株)国際電気通信基礎技術研究所,名古屋工業大学 産業戦略工学専攻)
松田 繁樹((独)情報通信研究機構)
杉浦 孔明((独)情報通信研究機構)
船越 孝太郎((株)ホンダ・リサーチ・インスティチュート・ジャパン)
中野 幹生((株)ホンダ・リサーチ・インスティチュート・ジャパン)
岡 夏樹(京都工芸繊維大学 大学院工芸科学研究科 情報工学部門)
時間06月09日(Wed) 15:35〜15:55
概要In this paper, we propose a novel method for a robot to detect robot-directed speech from other speech: to distinguish speech that users speak to a robot from speech that users speak to other people or to himself. The originality of this work is the introduction of Multimodal Semantic Confidence measure, which is used for domain classification
of input speech based on deciding whether the speech can be interpreted as a feasible action under the current physical situation in an object manipulation task. This measure is calculated by integrating speech, object, and motion confidence with weightings that are optimized by logistic regression. Then we integrate this measure with gaze
tracking, and conduct experiments under conditions of natural human-robot interactions. Experimental results show that the proposed method achieved a high performance of 94% and 96% in average recall and precision rates for robot-directed speech detection.
論文PDFファイル