You are on page 1of 3

N E U RA L N E T WO R KS

AI Overcomes Stumbling Block on Brain-Inspired Hardware

By A L L I S O N W H I T T E N

February 17, 2022

Algorithms that use the brain’s communication signal can now work on analog neuromorphic chips, which closely mimic our
energy-efficient brains.

The BrainScaleS-2 neuromorphic chip, developed by neuromorphic engineers at Heidelberg University,


uses tiny circuits that mimic the analog computing of the real neurons in our brains.

Heidelberg University

T
oday’s most successful artificial intelligence algorithms, artificial neural networks, are
loosely based on the intricate webs of real neural networks in our brains. But unlike our
highly efficient brains, running these algorithms on computers guzzles shocking amounts of
energy: The biggest models consume nearly as much power as five cars over their lifetimes.

Enter neuromorphic computing, a closer match to the design principles and physics of our brains that
could become the energy-saving future of AI. Instead of shuttling data over long distances between a
central processing unit and memory chips, neuromorphic designs imitate the architecture of the jelly-
like mass in our heads, with computing units (neurons) placed next to memory (stored in the synapses
that connect neurons). To make them even more brain-like, researchers combine neuromorphic chips
with analog computing, which can process continuous signals, just like real neurons. The resulting
chips are vastly different from the current architecture and computing mode of digital-only computers
that rely on binary signal processing of 0s and 1s.

With the brain as their guide, neuromorphic chips promise to one day demolish the energy
consumption of data-heavy computing tasks like AI. Unfortunately, AI algorithms haven’t played well
with the analog versions of these chips because of a problem known as device mismatch: On the chip,
tiny components within the analog neurons are mismatched in size due to the manufacturing process.
Because individual chips aren’t sophisticated enough to run the latest training procedures, the
algorithms must first be trained digitally on computers. But then, when the algorithms are transferred
to the chip, their performance breaks down once they encounter the mismatch on the analog hardware.

Now, a paper published last month in the Proceedings of the National Academy of Sciences has finally
revealed a way to bypass this problem. A team of researchers led by Friedemann Zenke at the Friedrich
Miescher Institute for Biomedical Research and Johannes Schemmel at Heidelberg University showed
that an AI algorithm known as a spiking neural network — which uses the distinctive communication
signal of the brain, known as a spike — could work with the chip to learn how to compensate for device
mismatch. The paper is a significant step toward analog neuromorphic computing with AI.

“The amazing thing is that it worked so well,” said Sander Bohte, a spiking neural network expert at
CWI, the national research institute for mathematics and computer science in the Netherlands. “It is
quite an achievement and likely a blueprint for more with analog neuromorphic systems.”

The importance of analog computing to brain-based computing is subtle. Digital computing can
effectively represent one binary aspect of the brain’s spike signal, an electrical impulse that shoots
through a neuron like a lightning bolt. As with a binary digital signal, either the spike is sent out, or it’s
not. But spikes are sent over time continuously — it’s an analog signal — and how our neurons decide
to send out a spike in the first place is also continuous, based on a voltage inside the cell that changes
over time. (When the voltage reaches a specific threshold compared to the voltage outside the cell, the
neuron sends a spike.)

“In analog lies the beauty of the brain’s core computations. Emulating this key aspect of the brain is
one of the main drivers of neuromorphic computing,” said Charlotte Frenkel, a neuromorphic
engineering researcher at the University of Zurich and ETH Zurich.

In 2011, a group of researchers at Heidelberg University began developing a neuromorphic chip with
both analog and digital aspects to closely model the brain for neuroscience experiments. Now led by
Schemmel, the team has unveiled the latest version of the chip, dubbed BrainScaleS-2. Every analog
neuron on the chip mimics a brain cell’s incoming and outgoing currents and voltage changes.

“You really have a dynamical system that is continuously exchanging information,” said Schemmel.
And because the materials have different electrical properties, the chip transfers information 1,000
times faster than our brains.

But since the analog neurons’ properties vary ever so slightly — the device mismatch problem — the
voltages and levels of current also varies across neurons. The algorithms can’t handle this, since they
were trained on computers with perfectly identical digital neurons, and suddenly their on-chip
performance plummets.

The new work shows a way forward. By including the chip in the training process, the authors showed
that spiking neural networks could learn how to correct for the varying voltages on the BrainScaleS-2
chip. “This training setup is one of the first convincing proofs that variability can not only be
compensated [for], but also likely exploited,” Frenkel said.

To deal with device mismatch, the team combined an approach that allows the chip to talk to the
computer with a new learning method called surrogate gradients, co-developed by Zenke specifically
for spiking neural networks. It works by changing the connections between neurons to minimize how
many errors a neural network makes in a task. (This is similar to the method used by non-spiking
neural networks, called backpropagation.)

Effectively, the surrogate gradient method was able to correct for the chip’s imperfections during
training on the computer. First, the spiking neural network performs a simple task using the varying
voltages of the analog neurons on the chip, sending recordings of the voltages back to the computer.
There, the algorithm automatically learns how to best alter connections between its neurons to still
play nicely with the analog neurons, continuously updating them on the chip while it learns. Then,
when the training is complete, the spiking neural network performs the task on the chip. The
researchers report that their network reached the same level of accuracy on a speech and vision task as
the top spiking neural networks performing the task on computers. In other words, the algorithm
learned exactly what changes it would need to make to overcome the device mismatch problem.

“The performance they achieved to do a real problem with this system is a big accomplishment,” said
Thomas Nowotny, a computational neuroscientist at the University of Sussex. And, as expected, they
do so with impressive energy efficiency; the authors say that running their algorithm on the chip
consumed about 1,000 times less energy than a standard processor would require.

However, Frenkel points out that while the energy consumption is good news so far, neuromorphic
chips will still need to prove themselves against hardware that has been optimized for similar speech
and vision recognition tasks, rather than standard processors. And Nowotny cautions that the
approach may have trouble scaling up to large practical tasks, since it still requires shuttling the data
back and forth between computer and chip.

The long-term goal is for spiking neural networks to train and run on neuromorphic chips from start
to finish, without any need for a computer at all. But that would require building a new chip generation,
which takes years, Nowotny said.

For now, Zenke and Schemmel’s team is happy to show that spiking neural network algorithms can
handle the minuscule variations between analog neurons on neuromorphic hardware. “You can rely on
60 or 70 years of experience and software history for digital computing,” Schemmel said. “For this
analog computing, we have to do everything on our own.”

You might also like