手势
计算机科学
对话框
串联(数学)
接口(物质)
人机交互
组分(热力学)
控制(管理)
构造(python库)
对话系统
语音识别
人工智能
万维网
程序设计语言
并行计算
气泡
最大气泡压力法
物理
组合数学
热力学
数学
作者
Toshiaki Nakagawa,Norihide Kitaoka
出处
期刊:Journal of the Acoustical Society of America
[Acoustical Society of America]
日期:2016-10-01
卷期号:140 (4_Supplement): 2963-2964
被引量:2
摘要
The recent development of autonomous vehicles has attracted much attention, but operating these vehicles may be too complex for average users. Therefore, we propose an intuitive, multimodal interface for the control of autonomous vehicles using speech and gesture recognition to interpret and execute the commands of users. For example, if the user says “turn there” while pointing at a landmark, the vehicle can utilize this behavior to correctly understand and comply with the user’s intent. To achieve this, we designed a two-part interface consisting of a multimodal understanding component and a dialog control component. Our multimodal understanding and dialog control components can be seen as a concatenation of two separate transducers. One transducer is used for multimodal understanding and the other for a conventional dialog system. We then construct a combined transducer from these two transducers. We developed various scenarios which might arise while operating an autonomous vehicle and displayed these scenes on a monitor. Subjects were then asked to operate a virtual car using speech commands and pointing gestures to control the vehicle while observing the monitor. The questionnaire results show that subjects felt they were able to easily and naturally operate the autonomous vehicle using utterances and gestures.
科研通智能强力驱动
Strongly Powered by AbleSci AI