DocumentCode
122912
Title
Accessible interfaces for robot assistants
Author
Lazewatsky, Daniel A. ; Smart, William D.
Author_Institution
Sch. of Mech., Ind., & Manuf. Eng., Oregon State Univ., Corvallis, OR, USA
fYear
2014
fDate
25-29 Aug. 2014
Firstpage
106
Lastpage
111
Abstract
Currently, high-level task control of robots is generally performed by using a graphical interface on a desktop or laptop computer. This type of mediated interaction is not natural, and can be problematic and cumbersome for persons with certain types of motor disabilities, and for people interacting with the robot when there are no computer displays present. In this work, we present a framework which enables the removal of such obvious intermediary devices and allows users to assign tasks to robots using interfaces embedded directly in the world, by projecting these interfaces directly onto surfaces and objects. We describe the implementation of the projected interface framework, and give several examples of tasks which can be performed with such an interface.
Keywords
graphical user interfaces; human-robot interaction; service robots; accessible interfaces; graphical interface; high-level robot task control; human-robot interaction; intermediary device removal; mediated interaction; motor disabled persons; robot assistants; robot task assignment; Context; Robot kinematics; Robot sensing systems; TV; Three-dimensional displays;
fLanguage
English
Publisher
ieee
Conference_Titel
Robot and Human Interactive Communication, 2014 RO-MAN: The 23rd IEEE International Symposium on
Conference_Location
Edinburgh
Print_ISBN
978-1-4799-6763-6
Type
conf
DOI
10.1109/ROMAN.2014.6926238
Filename
6926238
Link To Document