Autonomous, do-it-all robots are no longer the domain of science fiction or cartoons like 'The Jetsons', thanks to scientists from Brown University who have developed a robot that responds to human gestures.
The research team has demonstrated how a robot can follow nonverbal commands from a person in a variety of environments - indoors as well as outside - all without adjusting for lighting.
Advertisement"We have created a novel system where the robot will follow you at a precise distance, where you don't need to wear special clothing, you don't need to be in a special environment, and you don't need to look backward to track it," said Chad Jenkins, assistant professor of computer science at Brown University and the team's leader.
During the study, the researchers including Brown graduate students, showed how the robot responded as they used variety of hand-arm signals to instruct the robot, including "follow," "halt," "wait" and "door breach."
The Brown team started with a PackBot, a mechanized platform developed by iRobot that has been used widely by the U.S. military for bomb disposal, among other tasks.
They outfitted their robot with a commercial depth-imaging camera. They also geared the robot with a laptop that included novel computer programs that enabled the machine to recognize human gestures, decipher them and respond to them.
The researchers made two key advances with their robot. The first involved what scientists call visual recognition. They created a computer program, whereby the robot recognized a human by extracting a silhouette, as if a person were a virtual cutout.
This allowed the robot to home in on the human and receive commands without being distracted by other objects in the space.
The second advance involved the depth-imaging camera, where they used a CSEM Swiss Ranger, which uses infrared light to detect objects and to establish distances between the camera and the target object, and measure the distance between the camera and any other objects in the area.
The result is a robot that doesn't require remote control or constant vigilance, Jenkins said.
"What you really want is a robot that can act like a partner," Jenkins added. "You don't want to puppeteer the robot. You just want to supervise it, where you say, 'Here's your job. Now, go do it.'"
The study was presented at Human-Robot Interaction conference, in San Diego.
P Risk Factors Determining Asymmetrical Head Shapes In Infants Cycle Rally Marks World Kidney Day in Pune M
You May Also Like