Professional Documents
Culture Documents
Tao Hu1,2, Zaid J. Towfic1,3, Cengiz Pehlevan1, Alex Genkin1,4, and Dmitri B. Chklovskii1
1
INTRODUCTION
While the complexity of neuronal function is truly mindboggling, several of its salient features can be summarized in
the following diagram, Fig. 1A. A typical neuron receives
spike trains from thousands of upstream neurons via synapses
converting spikes into currents. The magnitude of each such
current is called a synaptic weight and its shape indicates lowpass (LP) filtering, or leaky integration. Synaptic currents
summate in the cell body, thus depolarizing the axonal
membrane to generate a spike train that then is transmitted to
thousands of downstream neurons. For the purposes of this
paper, we quantify the incoming and the outgoing spike trains
by a firing rate, which is the number of spikes per unit time.
The firing rate is a nonlinear function of the summed current
often approximated by a one-sided soft-threshold function.
The firing rate of a neuron typically follows a nonGaussian
distribution that is sparse (i.e. with a significant peak at zero)
and heavy-tailed (i.e. decaying slower than a Gaussian).
Finally, synaptic weights are continuously updated according
to the correlation between the activity of presynaptic and
postsynaptic neurons, known as Hebbian learning.
AVG Consulting
Brooklyn, NY
A. Offline setting
We start by postulating that a neuron computes a rank-1
representation of a temporally smoothed incoming data
M T
is temporally correlated on a timescale, , to improve signalto-noise ratio the representation error is convolved with a
smoothing
kernel
given
by
the
powers
of exp(1 ), 0 1 :
1 s x wy 2
T
t s t 2
s 0
w, y arg min
w ,y
t 1
2
2 y yt 2w1 w 1 w 2 w 2
arg min
w ,y
w yt
2
2
2 y yt 2 w 1 w 1 w 2 w
t 1
where xT 1
y = ST XT w , y w 22
2
w = ST X y , T w1
y 2 T w 2
ST f ,
0,
f , f
, yT arg min
w, y
, and
1
T 1
t 1
x t s w yt
s 0
xT s wy
2
2
2 y y
2
2
2T w1 w 1 T w 2 w
s 0
yT arg min 1
xT s w T 1 y
2
2
2 y y 1
s 0
1 s x w
T s T 1 2y
s 0
arg min
y
2
y
w T 1 2
w T 1
2
2
x
s
T s
, LP filtered, or leaky-integrated
s 0
presynaptic
inputs,
xT 1
x
s
s 0
T s
and
soft
2
2
xT xT 1 1 xT
yT ST w T 1 xT , y w T 1
III.
2
2
These equations are nothing else but a commonly used linearnonlinear or leaky integrate-and-fire model of a neuron, Fig.
1C. Note that by using recursion we avoided storing past data
keeping this step truly online.
Second, we fix the value of y and minimize the online loss
with respect to w:
wT
T
arg min
t 1
1 s x wy 2 2 w w 2
t s t 2 w1 1 w 2 2
s 0
2
y 1 x
s
arg min
w
t 1
s 0
T
yt
t s
2 w1 w 1 w 2 w
t 1
2
2
2
t
t 1
YT
xT xT 1 1 xT
2
yT ST w T 1 xT , y w T 1 2
2
YT YT 1 yT
u u y x u y Y
T 1
T
T
T 1 T
T
T
w T ST uT , T w1 YT T w 2
t
1. Algorithm selects value w
2. World reveals x t and yt
3. Algorithm suffers loss:
lt w t xt w t yt
2
2
2w1 w t
w 2 w t
2
2
R w 1 ,..., w T
lt w t min
t 1
l w
t
t 1
min
w
l (w ) ,
s 1
R w 1 ,..., w T 16 D w1 w 2 d 1 log T / w 2
2
Fig. 3. Leaky integration and soft thresholding in the online algorithm and
biological neurons. A) Leaky integration specified by the discounting factor.
B) Experimentally measured decay of a postsynaptic potential in response to a
presynpatic spike is similar to leaky integration [18]. C) Because of the l1norm regularizer sparse factorization algorithms apply soft threshold function
to the weighted sum of leaky integrated inputs. D) Experimentally measured
firing rate vs. injected current curve [19] is similar to the nonnegative half of
the soft threshold function.
[9]
[10]
[11]
[12]
[13]
Fig. 4. Experimentally observed nonGaussian distributions. Red: distribution
of firing rates (A) (Courtesy of J. Magee) and synaptic weights (B) [20] are
nonGaussian, sparse and heavy-tailed. Dashed blue: a Gaussian distribution
shown for comparison.
V.
CONCLUSION
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
ACKNOWLEDGMENT
[22]
[23]
REFERENCES
[1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[24]
[25]
[26]
[27]
[28]
[29]