Neuroscience

Articles and news from the latest research reports.

83 notes

This beer-pouring robot is programmed to anticipate human actions
A robot in Cornell’s Personal Robotics Lab has learned to foresee human action in order to step in and offer a helping hand, or more precisely, roll in and offer a helping claw.
Understanding when and where to pour a beer or knowing when to offer assistance opening a refrigerator door can be difficult for a robot because of the many variables it encounters while assessing the situation. Well, a team from Cornell has created a solution.
Gazing intently with a Microsoft Kinect 3-D camera and using a database of 3D videos, the Cornell robot identifies the activities it sees, considers what uses are possible with the objects in the scene and determines how those uses fit with the activities. It then generates a set of possible continuations into the future – such as eating, drinking, cleaning, putting away – and finally chooses the most probable. As the action continues, the robot constantly updates and refines its predictions.
"We extract the general principles of how people behave," said Ashutosh Saxena, Cornell professor of computer science and co-author of a new study tied to the research. "Drinking coffee is a big activity, but there are several parts to it." The robot builds a "vocabulary" of such small parts that it can put together in various ways to recognize a variety of big activities, he explained.
Saxena will join Cornell graduate student Hema S. Koppula as they present their research at the International Conference of Machine Learning, June 18-21 in Atlanta, and the Robotics: Science and Systems conference June 24-28 in Berlin, Germany.
In tests, the robot made correct predictions 82 percent of the time when looking one second into the future, 71 percent correct for three seconds and 57 percent correct for 10 seconds.
"Even though humans are predictable, they are only predictable part of the time," Saxena said. "The future would be to figure out how the robot plans its action. Right now we are almost hard-coding the responses, but there should be a way for the robot to learn how to respond."

This beer-pouring robot is programmed to anticipate human actions

A robot in Cornell’s Personal Robotics Lab has learned to foresee human action in order to step in and offer a helping hand, or more precisely, roll in and offer a helping claw.

Understanding when and where to pour a beer or knowing when to offer assistance opening a refrigerator door can be difficult for a robot because of the many variables it encounters while assessing the situation. Well, a team from Cornell has created a solution.

Gazing intently with a Microsoft Kinect 3-D camera and using a database of 3D videos, the Cornell robot identifies the activities it sees, considers what uses are possible with the objects in the scene and determines how those uses fit with the activities. It then generates a set of possible continuations into the future – such as eating, drinking, cleaning, putting away – and finally chooses the most probable. As the action continues, the robot constantly updates and refines its predictions.

"We extract the general principles of how people behave," said Ashutosh Saxena, Cornell professor of computer science and co-author of a new study tied to the research. "Drinking coffee is a big activity, but there are several parts to it." The robot builds a "vocabulary" of such small parts that it can put together in various ways to recognize a variety of big activities, he explained.

Saxena will join Cornell graduate student Hema S. Koppula as they present their research at the International Conference of Machine Learning, June 18-21 in Atlanta, and the Robotics: Science and Systems conference June 24-28 in Berlin, Germany.

In tests, the robot made correct predictions 82 percent of the time when looking one second into the future, 71 percent correct for three seconds and 57 percent correct for 10 seconds.

"Even though humans are predictable, they are only predictable part of the time," Saxena said. "The future would be to figure out how the robot plans its action. Right now we are almost hard-coding the responses, but there should be a way for the robot to learn how to respond."

Filed under robots robotics human action neuroscience technology science

  1. warlikeparakeet88 reblogged this from dermoosealini
  2. fuxkfuxkfuxk reblogged this from neurosciencestuff
  3. itsgracesdrunkmametown reblogged this from pluviaticus
  4. pluviaticus reblogged this from neurosciencestuff
  5. cardenvondraken reblogged this from neurosciencestuff
  6. muyfragil reblogged this from neurosciencestuff
  7. thygeekgoddess reblogged this from neurosciencestuff
  8. yannatosaurus reblogged this from neurosciencestuff and added:
    So we’re getting closer to a robot like Serge from Caprica! (Which I am currently watching and I don’t know how I forgot...
  9. dermoosealini reblogged this from neurosciencestuff
  10. rebelliving reblogged this from neurosciencestuff
  11. keatonaltom reblogged this from neurosciencestuff
  12. cavalokun reblogged this from neurosciencestuff
  13. laughingmad reblogged this from neurosciencestuff
  14. birdie-heichou reblogged this from neurosciencestuff
  15. thirteenthour reblogged this from neurosciencestuff
  16. fearisananchor reblogged this from neurosciencestuff
  17. browncrowns reblogged this from charmingpplincardigans
free counters