New intelligent algorithms could help robots to quickly recognize and respond to human gestures. Researchers at A*STAR Institute for Infocomm Research in Singapore have created a computer program which recognizes human gestures quickly and accurately, and requires very little training.
Many works of science fiction have imagined robots that could interact directly with people to provide entertainment, services or even health care. Robotics is now at a stage where some of these ideas can be realized, but it remains difficult to make robots easy to operate.
One option is to train robots to recognize and respond to human gestures. In practice, however, this is difficult because a simple gesture such as waving a hand may appear very different between different people. Designers must develop intelligent computer algorithms that can be 'trained' to identify general patterns of motion and relate them correctly to individual commands.
Now, Rui Yan and co-workers at the A*STAR Institute for Infocomm Research in Singapore have adapted a cognitive memory model called a localist attractor network (LAN) to develop a new system that recognize gestures quickly and accurately, and requires very little training.
"Since many social robots will be operated by non-expert users, it is essential for them to be equipped with natural interfaces for interaction with humans," says Yan. "Gestures are an obvious, natural means of human communication. Our LAN gesture recognition system only requires a small amount of training data, and avoids tedious training processes."
Yan and co-workers tested their software by integrating it with ShapeTape, a special jacket that uses fibre optics and inertial sensors to monitor the bending and twisting of hands and arms. They programmed the ShapeTape to provide data 80 times per second on the three-dimensional orientation of shoulders, elbows and wrists, and applied velocity thresholds to detect when gestures were starting.
In tests, five different users wore the ShapeTape jacket and used it to control a virtual robot through simple arm motions that represented commands such as forward, backwards, faster or slower. The researchers found that 99.15% of gestures were correctly translated by their system. It is also easy to add new commands, by demonstrating a new control gesture just a few times.
The next step in improving the gesture recognition system is to allow humans to control robots without the need to wear any special devices. Yan and co-workers are tackling this problem by replacing the ShapeTape jacket with motion-sensitive cameras.
"Currently we are building a new gesture recognition system by incorporating our method with a Microsoft Kinect camera," says Yan. "We will implement the proposed system on an autonomous robot to test its usability in the context of a realistic service task, such as cleaning!"
The above story is based on materials provided by The Agency for Science, Technology and Research (A*STAR). Note: Materials may be edited for content and length.
- Rui Yan, Keng Peng Tee, Yuanwei Chua, Haizhou Li, Huajin Tang. Gesture Recognition Based on Localist Attractor Networks with Application to Robot Control [Application Notes]. IEEE Computational Intelligence Magazine, 2012; 7 (1): 64 DOI: 10.1109/MCI.2011.2176767
Cite This Page: