Title :
How well can we encode spatial layout from sparse kinesthetic contact?
Author :
Klatzky, Roberta L. ; Lederman, Susan J.
Author_Institution :
Carnegie Mellon Univ., Pittsburgh, PA, USA
Abstract :
We investigated people´s ability to report the shape and scale of a spatial layout after sparse contact, without vision. We propose that the initial representation of sparsely contacted layout is kinesthetic. From this can be computed a configural representation that supports reports of shape and scale, but at the cost of increased error. In four experiments, participants´ fingers were guided to a two-point layout, after which they returned to the points or reported distance and/or angle, subject to a change in location and sometimes a rotation as well. Errors in reproducing inter-point distance, i.e., the scale of the layout, were smallest for the task of returning to the touched points and nearly twice as great when distance was reported at a new location. Errors in reproducing inter-point angle, i.e., the shape of the layout, were smallest for the task of returning to the touched points and nearly twice as great when angle was reported subject to rotation. The data highlight limitations on reporting the shape and scale of a haptically rendered layout after sparse contact.
Keywords :
haptic interfaces; human factors; interactive devices; errors; experiments; haptic interface; haptically rendered layout; inter-point angle; rotation; sparse contact; sparse kinesthetic contact; spatial layout; Costs; Fingers; Haptic interfaces; Keyboards; Lead; Mice; Muscles; Shape; Space stations; Tendons;
Conference_Titel :
Haptic Interfaces for Virtual Environment and Teleoperator Systems, 2003. HAPTICS 2003. Proceedings. 11th Symposium on
Print_ISBN :
0-7695-1890-7
DOI :
10.1109/HAPTIC.2003.1191269