ICRA 2012 Paper Abstract

Close

Paper TuC08.5

Takano, Wataru (University of Tokyo), Nakamura, Yoshihiko (University of Tokyo)

Bigram-Based Natural Language Model and Statistical Motion Symbol Model for Scalable Language of Humanoid Robots

Scheduled for presentation during the Regular Session "Human Detection and Tracking" (TuC08), Tuesday, May 15, 2012, 15:30−15:45, Meeting Room 8 (Wacipi)

2012 IEEE International Conference on Robotics and Automation, May 14-18, 2012, RiverCentre, Saint Paul, Minnesota, USA

This information is tentative and subject to change. Compiled on May 21, 2018

Keywords Recognition, Learning and Adaptive Systems, Gesture, Posture and Facial Expressions

Abstract

The language is a symbolic system unique to human being. The acquisition of language, which has its meanings in the real world, is important for robots to understand the environment and communicate with us in our daily life. This paper propose a novel approach to establish a fundamental framework for the robots which can understand language through their whole body motions. The proposed framework is composed of three modules : ``motion symbol", ``motion language model", and ``natural language model". In the motion symbol module, motion data is symbolized by Hidden Markov Models (HMMs). Each HMM represents abstract motion patterns. Then the HMMs are defined as motion symbols. The motion language model is stochastically designed for links between motion symbols and words. This model consists of three layers of motion symbols, latent variables and words. The connections between the motion symbol and the latent state, and between the latent state and the words is denoted by two kinds of probabilities respectively. One connection is represented by the probability that the motion symbol generates the latent state, and the other connection is represented by the probability that the latent state generates the words. Therefore, the motion language model can connect the motion symbols to the words through the latent state. The natural language model stochastically represents sequences of words. In this paper, a bigram, which is a special case of N-gram model, is adopted as the natura

 

 

Technical Content © IEEE Robotics & Automation Society

This site is protected by copyright and trademark laws under US and International law.
All rights reserved. © 2002-2018 PaperCept, Inc.
Page generated 2018-05-21  11:21:07 PST  Terms of use