News Release

A robot on EBRAINS has learned to combine vision and touch

Brain models are helping to discover better robot navigation methods

Business Announcement

Human Brain Project

Whiskeye Robot

image: The Whiskeye Robot view more 

Credit: HBP

How the brain lets us perceive and navigate the world is one of the most fascinating aspects of cognition. When orienting ourselves, we constantly combine information from all six senses in a seemingly effortless way - a feature that even the most advanced AI systems struggle to replicate.

On the new EBRAINS research infrastructure, cognitive neuroscientists, computational modellers and roboticists are now working together to shed new light on the neural mechanisms behind this, by creating robots whose internal workings mimic the brain.

"We believe that robots can be improved through the use of knowledge about the brain. But at the same time, this can also help us better understand the brain", says Cyriel Pennartz, a Professor of Cognition and Systems Neurosciences at the University of Amsterdam.

In the Human Brain Project, Pennartz has collaborated with computational modelers Shirin Dora, Sander Bohte and Jorge F. Mejias to create complex neural network architectures for perception based on real-life data from rats. Their model, dubbed "MultiPrednet" consists of modules for visual and tactile input, and a third that merges them.

"What we were able to replicate for the first time, is that the brain makes predictions across different senses", Pennartz explains. "So you can predict how something will feel from looking at it, and vice versa".

The way these networks "train" resembles how scientists think our brain learns: By constantly generating predictions about the world, comparing them to actual sensory inputs, and then adapting the network to avoid future error signals.

To test how the MultiPrednet performs in a body, the researchers teamed up with Martin Pearson at Bristol Robotics Lab. Together they integrated it into Whiskeye, a rodent-like robot that autonomously explores its environment, using head-mounted cameras for eyes, and 24 artificial whiskers to gather tactile information.

The researchers observed first indications that the brain-based model has an edge over traditional deep learning systems: Especially when it comes to navigation and recognition of familiar scenes, the MultiPredNet seems to perform better - a discovery the team now investigates further.

To accelerate this research, the robot has been recreated as a simulation on the Neurorobotics Platform of the EBRAINS research infrastructure. "This allows us to do long duration or even parallel experiments under controlled conditions", says Pearson. "We also plan to use the High Performance and Neuromorphic Computing Platforms for much more detailed models of control and perception in the future."

All code and analysis tools of the work are open on EBRAINS, so that researchers can run their own experiments. "It's a unique situation", says Pennartz: "We were able to say, here's an interesting model of perception based on neurobiology, and it would be great to test it on a larger scale with supercomputers and embodied in a robot. Doing this is normally very complicated, but EBRAINS makes it possible."

Katrin Amunts, Scientific Research Director of the HBP says: "To understand cognition, we will need to explore how the brain acts as part of the body in an environment. Cognitive neuroscience and robotics have much to gain from each other in this respect.The Human Brain Project brought these communities together, and now with our standing infrastructure, it's easier than ever to collaborate."

Pawel Swieboda, CEO of EBRAINS and Director General of the HBP, comments: "The robots of the future will benefit from innovations that connect insights from brain science to AI and robotics. With EBRAINS, Europe can be at the center of this shift to more bio-inspired AI and technology."

###

Video:

Perception and Recognition of Objects and Scenes
https://www.youtube.com/watch?v=0blgdAE1cbk

Learn how Human Brain Project scientists use EBRAINS to teach robots vision and touch

Video description on Youtube

Scientists in the Human Brain Project have constructed a brain-like, deep learning network that enables more efficient robot navigation. The so-called predictive coding networks merge information from visual and tactile sensors of a robot. Using EBRAINS, the brain-based network is connected to a virtual robotics platform for rapid testing and parallel experiments.

Access to the Experiment via the EBRAINS Knowledge Graph:

https://search.kg.ebrains.eu/instances/Model/2164c2b9bbb66b42ce358d108b5081ce


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.