Sensory Integration with Articulated Motion on a Humanoid Robot

This paper describes the integration of articulated motion with auditory and visual sensory information that enables a humanoid robot to achieve certain reflex actions that mimic those of people. Reflexes such as reach-and-grasp behavior enables the robot to learn, through experience, its own state...

Full description

Saved in:
Bibliographic Details
Main Authors: J. Rojas, R. A. Peters II
Format: Article
Language:English
Published: Wiley 2005-01-01
Series:Applied Bionics and Biomechanics
Online Access:http://dx.doi.org/10.1533/abbi.2004.0057
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1849307929431244800
author J. Rojas
R. A. Peters II
author_facet J. Rojas
R. A. Peters II
author_sort J. Rojas
collection DOAJ
description This paper describes the integration of articulated motion with auditory and visual sensory information that enables a humanoid robot to achieve certain reflex actions that mimic those of people. Reflexes such as reach-and-grasp behavior enables the robot to learn, through experience, its own state and that of the world. A humanoid robot with binaural audio input, stereo vision, and pneumatic arms and hands exhibited tightly coupled sensory-motor behaviors in four different demonstrations. The complexity of successive demonstrations was increased to show that the reflexive sensory-motor behaviors combine to perform increasingly complex tasks. The humanoid robot executed these tasks effectively and established the groundwork for the further development of hardware and software systems, sensory-motor vector-space representations, and coupling with higher-level cognition.
format Article
id doaj-art-1dba23e0841a4df992a75e82577f0a1c
institution Kabale University
issn 1176-2322
1754-2103
language English
publishDate 2005-01-01
publisher Wiley
record_format Article
series Applied Bionics and Biomechanics
spelling doaj-art-1dba23e0841a4df992a75e82577f0a1c2025-08-20T03:54:37ZengWileyApplied Bionics and Biomechanics1176-23221754-21032005-01-0123-417117810.1533/abbi.2004.0057Sensory Integration with Articulated Motion on a Humanoid RobotJ. Rojas0R. A. Peters II1Center for Intelligent Systems, Vanderbilt University, Nashville TN, USACenter for Intelligent Systems, Vanderbilt University, Nashville TN, USAThis paper describes the integration of articulated motion with auditory and visual sensory information that enables a humanoid robot to achieve certain reflex actions that mimic those of people. Reflexes such as reach-and-grasp behavior enables the robot to learn, through experience, its own state and that of the world. A humanoid robot with binaural audio input, stereo vision, and pneumatic arms and hands exhibited tightly coupled sensory-motor behaviors in four different demonstrations. The complexity of successive demonstrations was increased to show that the reflexive sensory-motor behaviors combine to perform increasingly complex tasks. The humanoid robot executed these tasks effectively and established the groundwork for the further development of hardware and software systems, sensory-motor vector-space representations, and coupling with higher-level cognition.http://dx.doi.org/10.1533/abbi.2004.0057
spellingShingle J. Rojas
R. A. Peters II
Sensory Integration with Articulated Motion on a Humanoid Robot
Applied Bionics and Biomechanics
title Sensory Integration with Articulated Motion on a Humanoid Robot
title_full Sensory Integration with Articulated Motion on a Humanoid Robot
title_fullStr Sensory Integration with Articulated Motion on a Humanoid Robot
title_full_unstemmed Sensory Integration with Articulated Motion on a Humanoid Robot
title_short Sensory Integration with Articulated Motion on a Humanoid Robot
title_sort sensory integration with articulated motion on a humanoid robot
url http://dx.doi.org/10.1533/abbi.2004.0057
work_keys_str_mv AT jrojas sensoryintegrationwitharticulatedmotiononahumanoidrobot
AT rapetersii sensoryintegrationwitharticulatedmotiononahumanoidrobot