Scientists have tapped neuromorphic computing to keep robots learning about new objects after they’ve been deployed.

For the uninitiated, neuromorphic computing replicates the neural structure of the human brain to create algorithms that can deal with the uncertainties of the natural world.

Intel Labs has developed one of the most notable architectures in the field: the Loihi neuromorphic chip.

Greetings, humanoids

Subscribe to our newsletter now for a weekly recap of our favorite AI stories in your inbox.

Loihi is comprised of around 130,000 artificial neurons, which send information to each other across a “spiking” neural network (SNN). The chips had already powered a range of systems, from an smart artificial skin to an electronic “nose” that recognizes scents emitted from explosives.

Intel Labs this week unveiled another application. The research unit teamed up with the Italian Institute of Technology and the Technical University of Munich to deploy Loihi in a new approach to continual learning for robotics.

Interactive learning

The method targets systems that interact with unconstrained environments, such as future robotic assistants for healthcare and manufacturing.

Existing deep neural networks can struggle with object learning in these scenarios, as they require extensive well-prepared training data — and careful retraining on new objects they encounter. The new neuromorphic approach aims to overcomes these limitations.

The researchers first implemented an SNN on Loihi. This architecture localizes learning to a single layer of plastic synapses. It also accounts for different views of objects by adding new neurons on demand.

As a result, the learning process unfolds autonomously while interacting with the user.

Neuromorphic simulations

The team tested their approach in a simulated 3D environment. In this setup, the robot actively senses objects by moving an event-based camera that functions as its eyes.

The camera’s sensor “sees” objects in a manner inspired by small fixational eye movements called “microsaccades.”  If the object it views is new, the SNN representation is learned or updated. If the object  is known, the network recognizes it and provides feedback to the user.

In a simulated setup, the robot actively senses objects by moving its eyes (event-based camera, or Dynamic Vision Sensor), generating “miscrosaccades”. The events collected are used to drive a spiking neural network (SNN) on the Loihi chip. If the object or the view is new, its SNN representation is learned or updated. If the object is known, it is recognized by the network and respective feedback is given to the user.
The data collected by the camera drives an SNN on the Loihi chip. Credit: Intel Labs
In a simulated setup, the robot actively senses objects by moving its eyes (event-based camera, or Dynamic Vision Sensor), generating “miscrosaccades”. The events collected are used to drive a spiking neural network (SNN) on the Loihi chip. If the object or the view is new, its SNN representation is learned or updated. If the object is known, it is recognized by the network and respective feedback is given to the user.

The team say their method required up to 175-times lower energy to provide similar or better speed and accuracy than conventional methods running on a CPU.

They now need to test their algorithm in the real-world with actual robots.

“Our goal is to apply similar capabilities to future robots that work in interactive settings, enabling them to adapt to the unforeseen and work more naturally alongside humans,” Yulia Sandamirskaya, the study’s senior author, said in a statement. 

Their study, which was named “Best Paper” at this year’s International Conference on Neuromorphic Systems (ICONS), can be read here.

Advantages of overseas domestic helper.