• DocumentCode
    2017114
  • Title

    Modeling and composing gestures for human-robot interaction

  • Author

    Tay, Junyun ; Veloso, Manuela

  • Author_Institution
    Dept. of Mech. Eng., Carnegie Mellon Univ., Pittsburgh, PA, USA
  • fYear
    2012
  • fDate
    9-13 Sept. 2012
  • Firstpage
    107
  • Lastpage
    112
  • Abstract
    We formalize the representation of gestures and present a model that is capable of synchronizing expressive and relevant gestures with text-to-speech input. A gesture consists of gesture primitives that are executed simultaneously. We formally define the gesture primitive and introduce the concept of a spatially targeted gesture primitive, i.e., a gesture primitive that is directed at a target of interest. The spatially targeted gesture primitive is useful for situations where the direction of the gesture is important for meaningful human-robot interaction. We contribute an algorithm to determine how a spatially targeted gesture primitive is generated. We also contribute a process to analyze the input text, determine relevant gesture primitives from the input text, compose gestures from gesture primitives and rank the combinations of gestures. We propose a set of criteria that weights and ranks the combinations of gestures. Although we illustrate the utility of our model, algorithm and process using a NAO humanoid robot, our contributions are applicable to other robots.
  • Keywords
    human-robot interaction; humanoid robots; speech synthesis; NAO humanoid robot; gesture composition; gesture modeling; human-robot interaction; spatially targeted gesture primitive; text-to-speech input; Collision avoidance; Joints; Robot kinematics; Speech; Timing; Vectors;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    RO-MAN, 2012 IEEE
  • Conference_Location
    Paris
  • ISSN
    1944-9445
  • Print_ISBN
    978-1-4673-4604-7
  • Electronic_ISBN
    1944-9445
  • Type

    conf

  • DOI
    10.1109/ROMAN.2012.6343739
  • Filename
    6343739