Skip to main content

14 Robots Teach Themselves to Walk Into a Bar

Can robots be trained to pick up objects they've never seen before?  Recent research indicates that the best approach may be to allow robots to experiment on their own, and then share that knowledge with other robots.  

How To Train Your Robot

Until now, robots were programmed to recognize specific objects and to relate to those objects in a pre-programmed fashion. This is a good approach for predictable tasks, in planned locations, with previously experienced objects. In the real world, life is what happens while you’re making other plans.
At Google, Sergey Levine and his team took fourteen robotic arms, networked them together, and used (CNN) convolutional neural networks to let these robots learn on their own how to pick up small objects including a cup, tape dispenser, and a toy dolphin.
Over time, and nearly one million practice runs, the robots were able to self-correct and optimize their actions.  They learned to pick up objects faster and more often, employing  self-taught strategies - such as pushing an object away in order to reach another, and developed different techniques to pick up soft versus hard objects.

The robots learned from practice and sharing knowledge with one another, and without human interaction or pre-programming. In fact, the only human contribution was to replace the objects in the trays. 

During the study, the robots were positioned in front of trays with randomly placed objects varying in shape, scale, weight and hardness. Each of the 14 robot arms had a similar, but not identical design. 
Each robot’s successes and failures were fed to a shared convoluted neural network, a “robo-wiki” of shared knowledge accessible to the 14 robots. 
“Our experimental evaluation demonstrates that our method achieves effective real-time control, can successfully grasp novel objects, and corrects mistakes by continuous robotic hand-eye coordination...”  
Google’s project builds on the concept that robots can teach other robots how to perform specific tasks—or, more accurately, how one artificial intelligence can help another AI improve, without a human ever intervening. 
Here's a link to the Google Research: Learning Hand-Eye Coordination for Robotic Grasping with Deep Learning and Large-Scale Data Collection - Sergey Levine, Peter Pastor, Alex Krizhevsky, and Deirdre Quillen

The robots taught each other to pick up hard and soft objects in different ways, without human guidance. 

For an object that was perceived as rigid the robots would just grasp the outer edges and squeeze to hold it tightly. But for a soft object, like a sponge or the dolphin, the robots "learned" that it was easier to place one gripping "finger" in the middle and one around the edge, and then just squeeze.
The good news -- the Robots still needed humans to place objects in the trays.
Brown University project called “RoboBrain,” plans to store all machine-learned behaviors into a knowledge database which other bots can tap into. The project has been nicknamed “the world’s first knowledge engine for robots.” Some people call in "SkyNet."
Robo Brain is a large-scale computational system that learns from publicly available Internet resources, computer simulations, and real-life robot trials.
It accumulates everything robotics into a comprehensive and interconnected knowledge base. Applications include prototyping for robotics research, household robots, and self-driving cars. 

The Claw Is Our Master: He chooses who will go and who will stay.
Regarding robots walking into a bar... you would think that 14 robots using convolutional neural networks would be great at telling jokes. They're not.
The best robot-written humor I've found is derivative of human jokes. For example, the CMU bot, dubbed Data, began a recent robot standup session with,

“A doctor says to his patient, ‘I have bad news and worse news. The bad news is that you only have 24 hours to live.'” The patient wondered how the news could possibly get worse. “I’ve been robo-dialing you since yesterday...”
More on robots, AI and humor, here: It's hard work being funny--especially for robots.
Click this link for more information regarding Deep Learning and  Convoluted Neural Networks.
(c) David J. Katz, 2016 - New York City

Popular posts from this blog

Warning, Car Porn

The signature feature is the Rolls Royce Wraith’s Starlight Headliner, consisting of 1,340 LEDs hand-sewn to create an effect of owning one’s personal night sky filled with stars...

Warning, content below represents a man's libidinous fascination with an automobile. It is not Lolita; after all Bradley Berman, the author, is not Nabokov and the Wraith is not underaged. Nonetheless, I find myself simultaneously repulsed... and seduced. David J. Katz

Annotated Guide To Men's Belts

The Complete Guide To Men’s BeltsArticle By  on 11th March 2014 | @gabrielweil


Discounts, Discovery & Delight: 3Ds for Retail Success

In fashion and retail, Dopamine is the drug of choice. Technically, Dopamine is the neurotransmitter of “desire.” Dopamine leaps across synapses in our brain to control our reward and pleasure centers. It enables craving. It induces repeat behaviors. It makes us want more. Therefore, it is in our best interest to create products and experiences which induce the release of dopamine in our consumers. We could use some dopamine for ourselves, too. In our fashion and retail world, there are three primary stimuli, "3Ds," we can control to deliver hits of dopamine: Discounts, Discovery and Delight.