"Robots Can Learn a Range of Visual Object Manipulation Skills Entirely on Their Own"
Berkeley researchers have programmed the robot, Vestri, to complete tasks like a baby would – by playing with objects and then imagining how to get the task done. (UC Berkeley video by Roxanne Makasdjian and Stephen McNally)
UC Berkeley researchers have developed a robotic learning technology that enables robots to imagine the future of their actions so they can figure out how to manipulate objects they have never encountered before. In the future, this technology could help self-driving cars anticipate future events on the road and produce more intelligent robotic assistants in homes, but the initial prototype focuses on learning simple manual skills entirely from autonomous play.
Using this technology, called visual foresight, the robots can predict what their cameras will see if they perform a particular sequence of movements. These robotic imaginations are still relatively simple for now – predictions made only several seconds into the future – but they are enough for the robot to figure out how to move objects around on a table without disturbing obstacles. Crucially, the robot can learn to perform these tasks without any help from humans or prior knowledge about physics, its environment or what the objects are. That’s because the visual imagination is learned entirely from scratch from unattended and unsupervised exploration, where the robot plays with objects on a table. After this play phase, the robot builds a predictive model of the world, and can use this model to manipulate new objects that it has not seen before.
"In the same way that we can imagine how our actions will move the objects in our environment, this method can enable a robot to visualize how different behaviors will affect the world around it," said Sergey Levine, assistant professor in Berkeley's Department of Electrical Engineering and Computer Sciences, whose lab developed the technology. "This can enable intelligent planning of highly flexible skills in complex real-world situations."
The research team [was scheduled] to perform a demonstration of the visual foresight technology at the Neural Information Processing Systems conference in Long Beach, California, on December 5.
At the core of this system is a deep learning technology based on convolutional recurrent video prediction, or dynamic neural advection (DNA). DNA-based models predict how pixels in an image will move from one frame to the next based on the robot’s actions. Recent improvements to this class of models, as well as greatly improved planning capabilities, have enabled robotic control based on video prediction to perform increasingly complex tasks, such as sliding toys around obstacles and repositioning multiple objects.
Pages: 1 · 2
More Articles
- Berkeley Talks: Jessica Morse On How We Can Live With Fire
- Julia Sneden Wrote: Love Your Library
- Scientific Energy Breakeven: Advancements in National Defense and the Future of Clean Power
- "Henry Ford Innovation Nation", a Favorite Television Show
- Indoor and Vertical Farming May Be Part of the Solution to Rising Demands for Food and Limited Natural Resources
- National Institutes of Health: COVID-19 Vaccines Linked to Small Increase in Menstrual Cycle Length
- Ask KHN (Kaiser Health News) - PolitiFact: Is My Cloth Mask Good Enough? The 2022 Edition
- Rose Madeline Mula Writes: I’ve Got A Secret – NOT!
- On A Chilly Saturday, Winter Graduates Turn to Their Future: “Some of (your) most important lessons came from a real-life curriculum no one ever anticipated”
- The Stanford Center on Longevity: The New Map of Life