Professional Documents
Culture Documents
Presentation 31 01 2022
Presentation 31 01 2022
applications to neurosciences
1
ex-PhD student at I3S laboratory, Université Côte d’Azur
Articles:
Simulation of stochastic processes (Ornstein-Uhlenbeck)[1] (published)
Simulation of large networks of point processes[2] (in-press)
Application to large networks of point processes[3] (submitted)
Software:
GODDESS: generic discrete-event simulator
SPIKES: point process simulation
RECONSTRUCTION: causal interaction estimation
Supervisor: GUI for SPIKES and RECONSTRUCTION
Articles:
Simulation of stochastic processes (Ornstein-Uhlenbeck)[1] (published)
Simulation of large networks of point processes[2] (in-press)
Application to large networks of point processes[3] (submitted)
Software:
GODDESS: generic discrete-event simulator
SPIKES: point process simulation
RECONSTRUCTION: causal interaction estimation
Supervisor: GUI for SPIKES and RECONSTRUCTION
5/39
Context: neuroscience
Spikes (= points)
5/39
Context: neuroscience
Spikes (= points)
Spike
Sorting
5/39
Context: neuroscience
Spikes (= points)
Spike
Sorting
5/39
Point processes
6/39
Point processes
𝑡
𝑇1 𝑇2 𝑇3 𝑇4 𝑇5 𝑇6
6/39
Point processes
𝑡
𝑇1 𝑇2 𝑇3 𝑇4 𝑇5 𝑇6
Conditional Intensity
Some family described by conditional intensity 𝜙𝑡 (𝑥𝑡 ) (≈ rate):
6/39
Point processes
𝑡
𝑇1 𝑇2 𝑇3 𝑇4 𝑇5 𝑇6
Conditional Intensity
Some family described by conditional intensity 𝜙𝑡 (𝑥𝑡 ) (≈ rate):
6/39
Example: homogeneous Poisson point process
7/39
Example: homogeneous Poisson point process
𝜙𝑡 (𝑥𝑡 ) = Γ
𝑡
𝑇1 𝑇2 𝑇3 𝑇4 𝑇5 𝑇6
7/39
Example: homogeneous Poisson point process
𝜙𝑡 (𝑥𝑡 ) = Γ
𝑡
𝑇1 𝑇2 𝑇3 𝑇4 𝑇5 𝑇6
ℰ (Γ) ℰ (Γ)
Property
Interpoint intervals are independent and exponentially distributed
7/39
Why use point processes in neuroscience?
10/39
Interactions and time asynchrony
𝑗
𝜙𝑖,𝑡 (𝑥𝑡 ) ℎ𝑗𝑖 (𝑡)
10/39
Interactions and time asynchrony
𝑗
𝜙𝑖,𝑡 (𝑥𝑡 ) ℎ𝑗𝑖 (𝑡)
10/39
Interactions and time asynchrony
𝑗
𝜙𝑖,𝑡 (𝑥𝑡 ) ℎ𝑗𝑖 (𝑡)
10/39
Interactions and time asynchrony
𝑗
𝜙𝑖,𝑡 (𝑥𝑡 ) ℎ𝑗𝑖 (𝑡)
10/39
Interactions and time asynchrony
𝑗
𝜙𝑖,𝑡 (𝑥𝑡 ) ℎ𝑗𝑖 (𝑡)
Time asynchrony
At any time 𝑡, there can be only 1 point among all the processes.
10/39
Interactions and time asynchrony
𝑗
𝜙𝑖,𝑡 (𝑥𝑡 ) ℎ𝑗𝑖 (𝑡)
Time asynchrony
At any time 𝑡, there can be only 1 point among all the processes. This
is a given when the conditional intensity exists.
10/39
Multivariate Hawkes point process
𝑀 𝑡
ℎ (𝑡 − 𝑠) d𝑥𝑠
𝑖 𝑗
𝜙𝑖,𝑡 (𝑥𝑡 ) = 𝑓 𝜇 + ∑
( ∫ 𝑗𝑖 )
𝑗=1 0
11/39
Multivariate Hawkes point process
𝑀
𝜙𝑖,𝑡 (𝑥𝑡 ) = 𝜇 𝑖 + ∑ ∑ ℎ𝑗𝑖 (𝑡 − 𝑇𝑗 )
𝑗=1 𝑗
𝑇𝑗 <𝑡,𝑇𝑗 ∈𝑥𝑡
11/39
Multivariate Hawkes point process
𝑀
𝜙𝑖,𝑡 (𝑥𝑡 ) = 𝜇 𝑖 + ∑ ∑ ℎ𝑗𝑖 (𝑡 − 𝑇𝑗 )
𝑗=1 𝑗
𝑇𝑗 <𝑡,𝑇𝑗 ∈𝑥𝑡
11/39
Multivariate Hawkes point process
𝑀
𝜙𝑖,𝑡 (𝑥𝑡 ) = 𝜇 𝑖 + ∑ ∑ ℎ𝑗𝑖 (𝑡 − 𝑇𝑗 )
𝑗=1 𝑇 <𝑡,𝑇 ∈𝑥𝑗
𝑗 𝑗 𝑡
𝜙𝑖,𝑡 (𝑥𝑡 )
𝜇𝑖
𝑖
11/39
Multivariate Hawkes point process
𝑀
𝜙𝑖,𝑡 (𝑥𝑡 ) = 𝜇 𝑖 + ∑ ∑ ℎ𝑗𝑖 (𝑡 − 𝑇𝑗 )
𝑗=1 𝑇 <𝑡,𝑇 ∈𝑥𝑗
𝑗 𝑗 𝑡
𝑗
𝑇𝑗
𝜙𝑖,𝑡 (𝑥𝑡 )
𝜇𝑖
𝑖
11/39
Multivariate Hawkes point process
𝑀
𝜙𝑖,𝑡 (𝑥𝑡 ) = 𝜇 𝑖 + ∑ ∑ ℎ𝑗𝑖 (𝑡 − 𝑇𝑗 )
𝑗=1 𝑇 <𝑡,𝑇 ∈𝑥𝑗
𝑗 𝑗 𝑡
𝑗
𝑇𝑗
ℎ𝑗𝑖 (𝑡)
𝜙𝑖,𝑡 (𝑥𝑡 )
𝜇𝑖
𝑖
11/39
Multivariate Hawkes point process
𝑀
𝜙𝑖,𝑡 (𝑥𝑡 ) = 𝜇 𝑖 + ∑ ∑ ℎ𝑗𝑖 (𝑡 − 𝑇𝑗 )
𝑗=1 𝑇 <𝑡,𝑇 ∈𝑥𝑗
𝑗 𝑗 𝑡
𝑗
𝑇𝑗
ℎ𝑗𝑖 (𝑡)
𝜙𝑖,𝑡 (𝑥𝑡 )
𝑖 𝑇𝑗
11/39
Thinning
𝜙𝑡 (𝑥𝑡 )
12/39
Thinning
𝜙𝑡 (𝑥𝑡 )
Γ
12/39
Thinning
𝜙𝑡 (𝑥𝑡 )
Γ
𝑡
ℰ (Γ) ℰ (Γ) ℰ (Γ)
12/39
Thinning
𝜙𝑡 (𝑥𝑡 )
Γ
𝑇1
𝑡
𝜙𝑇1 (𝑥𝑇1 )
Accepted, P ∼ Γ
12/39
Thinning
𝜙𝑡 (𝑥𝑡 )
Γ
𝑇2
𝑡
𝜙𝑇2 (𝑥𝑇2 )
Rejected, P ∼ 1 − Γ
12/39
Thinning
𝜙𝑡 (𝑥𝑡 )
Γ
𝑇3
𝑡
𝜙𝑇3 (𝑥𝑇3 )
Accepted, P ∼ Γ
12/39
Time-rescaling theorem[6]
6 E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, and L. M. Frank. “The time-rescaling theorem and
its application to neural spike train data analysis”. In: Neural computation 14.2 (2002), pp. 325–346.
13/39
Time-rescaling theorem[6]
6 E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, and L. M. Frank. “The time-rescaling theorem and
its application to neural spike train data analysis”. In: Neural computation 14.2 (2002), pp. 325–346.
13/39
Time-rescaling theorem[6]
{Φ𝑇𝑘 (𝑥𝑇𝑘 ) / 𝑇𝑘 ∈ 𝑥𝑇 }
6 E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, and L. M. Frank. “The time-rescaling theorem and
its application to neural spike train data analysis”. In: Neural computation 14.2 (2002), pp. 325–346.
13/39
Time-rescaling theorem[6]
Time-rescaling theorem
∀(𝑇𝑘 < 𝑇𝑘+1 ) ∈ 𝑥𝑇 , Φ𝑇𝑘+1 (𝑥𝑇𝑘+1 ) − Φ𝑇𝑘 (𝑥𝑇𝑘 ) ∼ ℰ (1)
6 E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, and L. M. Frank. “The time-rescaling theorem and
its application to neural spike train data analysis”. In: Neural computation 14.2 (2002), pp. 325–346.
13/39
Time-rescaling theorem[6]
Time-rescaling theorem
∀(𝑇𝑘 < 𝑇𝑘+1 ) ∈ 𝑥𝑇 , Φ𝑇𝑘+1 (𝑥𝑇𝑘+1 ) − Φ𝑇𝑘 (𝑥𝑇𝑘 ) ∼ − log (𝒰(0, 1])
𝜙𝑡 (𝑥𝑡 )
ℰ (1)
6 E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, and L. M. Frank. “The time-rescaling theorem and
its application to neural spike train data analysis”. In: Neural computation 14.2 (2002), pp. 325–346.
13/39
Time-rescaling theorem[6]
Time-rescaling theorem
∀(𝑇𝑘 < 𝑇𝑘+1 ) ∈ 𝑥𝑇 , Φ𝑇𝑘+1 (𝑥𝑇𝑘+1 ) − Φ𝑇𝑘 (𝑥𝑇𝑘 ) ∼ − log (𝒰(0, 1])
𝜙𝑡 (𝑥𝑡 )
ℰ (1)
6 E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, and L. M. Frank. “The time-rescaling theorem and
its application to neural spike train data analysis”. In: Neural computation 14.2 (2002), pp. 325–346.
13/39
Time-rescaling theorem[6]
Time-rescaling theorem
∀(𝑇𝑘 < 𝑇𝑘+1 ) ∈ 𝑥𝑇 , Φ𝑇𝑘+1 (𝑥𝑇𝑘+1 ) − Φ𝑇𝑘 (𝑥𝑇𝑘 ) ∼ − log (𝒰(0, 1])
𝜙𝑡 (𝑥𝑡 )
ℰ (1)
6 E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, and L. M. Frank. “The time-rescaling theorem and
its application to neural spike train data analysis”. In: Neural computation 14.2 (2002), pp. 325–346.
13/39
Time-rescaling theorem[6]
Time-rescaling theorem
∀(𝑇𝑘 < 𝑇𝑘+1 ) ∈ 𝑥𝑇 , Φ𝑇𝑘+1 (𝑥𝑇𝑘+1 ) − Φ𝑇𝑘 (𝑥𝑇𝑘 ) ∼ − log (𝒰(0, 1])
𝜙𝑡 (𝑥𝑡 )
ℰ (1)
6 E. N. Brown, R. Barbieri, V. Ventura, R. E. Kass, and L. M. Frank. “The time-rescaling theorem and
its application to neural spike train data analysis”. In: Neural computation 14.2 (2002), pp. 325–346.
13/39
The Gillespie-inspired methods
A galaxy of methods
Doob’s algorithm (Markov processes)[7] .
7 J. L. Doob. “Topics in the theory of Markoff chains”. In: Transactions of the American Mathematical
Society 52.1 (1942), pp. 37–64.
8 D. T. Gillespie. “A general method for numerically simulating the stochastic time evolution of coupled
chemical reactions”. In: Journal of Computational Physics 22.4 (1976), pp. 403–434. issn: 0021-9991.
9 A. Bouchard-Côté, S. J. Vollmer, and A. Doucet. “The bouncy particle sampler: A nonreversible
rejection-free Markov chain Monte Carlo method”. In: Journal of the American Statistical Association
113.522 (2018), pp. 855–867.
10 P. A. W. Lewis and G. S. Shedler. “Simulation of nonhomogeneous poisson processes by thinning”. In:
Naval Research Logistics Quarterly 26.3 (1979), pp. 403–413. eprint:
https://onlinelibrary.wiley.com/doi/pdf/10.1002/nav.3800260304.
11 J. Møller and J. G. Rasmussen. “Perfect simulation of Hawkes processes”. In: Advances in applied
probability 37.3 (2005), pp. 629–646.
12 E. Bacry, I. Mastromatteo, and J.-F. Muzy. “Hawkes processes in finance”. In: Market Microstructure
and Liquidity 1.01 (2015), p. 1550005.
14/39
The Gillespie-inspired methods
A galaxy of methods
Doob’s algorithm (Markov processes)[7] .
Popularized by Gillespie, rediscovered by physicists[8,9]
7 J. L. Doob. “Topics in the theory of Markoff chains”. In: Transactions of the American Mathematical
Society 52.1 (1942), pp. 37–64.
8 D. T. Gillespie. “A general method for numerically simulating the stochastic time evolution of coupled
chemical reactions”. In: Journal of Computational Physics 22.4 (1976), pp. 403–434. issn: 0021-9991.
9 A. Bouchard-Côté, S. J. Vollmer, and A. Doucet. “The bouncy particle sampler: A nonreversible
rejection-free Markov chain Monte Carlo method”. In: Journal of the American Statistical Association
113.522 (2018), pp. 855–867.
10 P. A. W. Lewis and G. S. Shedler. “Simulation of nonhomogeneous poisson processes by thinning”. In:
Naval Research Logistics Quarterly 26.3 (1979), pp. 403–413. eprint:
https://onlinelibrary.wiley.com/doi/pdf/10.1002/nav.3800260304.
11 J. Møller and J. G. Rasmussen. “Perfect simulation of Hawkes processes”. In: Advances in applied
probability 37.3 (2005), pp. 629–646.
12 E. Bacry, I. Mastromatteo, and J.-F. Muzy. “Hawkes processes in finance”. In: Market Microstructure
and Liquidity 1.01 (2015), p. 1550005.
14/39
The Gillespie-inspired methods
A galaxy of methods
Doob’s algorithm (Markov processes)[7] .
Popularized by Gillespie, rediscovered by physicists[8,9]
Thinning method by Lewis and Schedler[10]
7 J. L. Doob. “Topics in the theory of Markoff chains”. In: Transactions of the American Mathematical
Society 52.1 (1942), pp. 37–64.
8 D. T. Gillespie. “A general method for numerically simulating the stochastic time evolution of coupled
chemical reactions”. In: Journal of Computational Physics 22.4 (1976), pp. 403–434. issn: 0021-9991.
9 A. Bouchard-Côté, S. J. Vollmer, and A. Doucet. “The bouncy particle sampler: A nonreversible
rejection-free Markov chain Monte Carlo method”. In: Journal of the American Statistical Association
113.522 (2018), pp. 855–867.
10 P. A. W. Lewis and G. S. Shedler. “Simulation of nonhomogeneous poisson processes by thinning”. In:
Naval Research Logistics Quarterly 26.3 (1979), pp. 403–413. eprint:
https://onlinelibrary.wiley.com/doi/pdf/10.1002/nav.3800260304.
11 J. Møller and J. G. Rasmussen. “Perfect simulation of Hawkes processes”. In: Advances in applied
probability 37.3 (2005), pp. 629–646.
12 E. Bacry, I. Mastromatteo, and J.-F. Muzy. “Hawkes processes in finance”. In: Market Microstructure
and Liquidity 1.01 (2015), p. 1550005.
14/39
The Gillespie-inspired methods
A galaxy of methods
Doob’s algorithm (Markov processes)[7] .
Popularized by Gillespie, rediscovered by physicists[8,9]
Thinning method by Lewis and Schedler[10]
Peculiar methods: immigrant-birth (a.k.a. clustering)[11,12]
7 J. L. Doob. “Topics in the theory of Markoff chains”. In: Transactions of the American Mathematical
Society 52.1 (1942), pp. 37–64.
8 D. T. Gillespie. “A general method for numerically simulating the stochastic time evolution of coupled
chemical reactions”. In: Journal of Computational Physics 22.4 (1976), pp. 403–434. issn: 0021-9991.
9 A. Bouchard-Côté, S. J. Vollmer, and A. Doucet. “The bouncy particle sampler: A nonreversible
rejection-free Markov chain Monte Carlo method”. In: Journal of the American Statistical Association
113.522 (2018), pp. 855–867.
10 P. A. W. Lewis and G. S. Shedler. “Simulation of nonhomogeneous poisson processes by thinning”. In:
Naval Research Logistics Quarterly 26.3 (1979), pp. 403–413. eprint:
https://onlinelibrary.wiley.com/doi/pdf/10.1002/nav.3800260304.
11 J. Møller and J. G. Rasmussen. “Perfect simulation of Hawkes processes”. In: Advances in applied
probability 37.3 (2005), pp. 629–646.
12 E. Bacry, I. Mastromatteo, and J.-F. Muzy. “Hawkes processes in finance”. In: Market Microstructure
and Liquidity 1.01 (2015), p. 1550005.
14/39
Ogata’s algorithm[13] (next occuring point)
13 Y. Ogata. “On Lewis’ simulation method for point processes”. In: IEEE transactions on information
theory 27.1 (1981), pp. 23–31.
15/39
Ogata’s algorithm[13] (next occuring point)
𝜙1,𝑡 (𝑥𝑡 )
𝜙1,𝑡 (𝑥𝑡 )
13 Y. Ogata. “On Lewis’ simulation method for point processes”. In: IEEE transactions on information
theory 27.1 (1981), pp. 23–31.
15/39
Ogata’s algorithm[13] (next occuring point)
𝜙1,𝑡 (𝑥𝑡 )
𝜙2,𝑡 (𝑥𝑡 )
1 2
1+2
13 Y. Ogata. “On Lewis’ simulation method for point processes”. In: IEEE transactions on information
theory 27.1 (1981), pp. 23–31.
15/39
Ogata’s algorithm[13] (next occuring point)
𝜙1,𝑡 (𝑥𝑡 )
𝜙𝑀,𝑡 (𝑥𝑡 )
1+2
⋮ ⋮ ⋮
𝜙1,𝑡 (𝑥𝑡 ) + ⋯ + 𝜙𝑀,𝑡 (𝑥𝑡 )
joint process
13 Y. Ogata. “On Lewis’ simulation method for point processes”. In: IEEE transactions on information
theory 27.1 (1981), pp. 23–31.
15/39
Ogata’s algorithm[13] (next occuring point)
𝜙1,𝑡 (𝑥𝑡 )
1+2
⋮ ⋮ ⋮
𝜙1,𝑡 (𝑥𝑡 ) + ⋯ + 𝜙𝑀,𝑡 (𝑥𝑡 )
joint process
13 Y. Ogata. “On Lewis’ simulation method for point processes”. In: IEEE transactions on information
theory 27.1 (1981), pp. 23–31.
15/39
Ogata’s algorithm[13] (next occuring point)
𝜙1,𝑡 (𝑥𝑡 )
1+2
⋮ ⋮ ⋮
𝜙1,𝑡 (𝑥𝑡 ) + ⋯ + 𝜙𝑀,𝑡 (𝑥𝑡 )
joint process
𝑇
13 Y. Ogata. “On Lewis’ simulation method for point processes”. In: IEEE transactions on information
theory 27.1 (1981), pp. 23–31.
15/39
Ogata’s algorithm[13] (next occuring point)
5 6
8 9
13 Y. Ogata. “On Lewis’ simulation method for point processes”. In: IEEE transactions on information
theory 27.1 (1981), pp. 23–31.
15/39
Ogata’s algorithm[13] (next occuring point)
𝜙𝑖,𝑡 (𝑥𝑡 ) = 𝜇 𝑖 + ∑𝑀
𝑗=1 ∑𝑇 <𝑡,𝑇 ∈𝑥 ℎ𝑗𝑖 (𝑡 − 𝑇𝑗 )
𝑗
𝑗 𝑗 𝑡
8 9
13 Y. Ogata. “On Lewis’ simulation method for point processes”. In: IEEE transactions on information
theory 27.1 (1981), pp. 23–31.
15/39
16/39
Bottleneck
17/39
Using context
Neuroscience:
Sparse graph (synaptic connectivity is big, but neuronal
connectivity is very low)
18/39
Using context
Neuroscience:
Sparse graph (synaptic connectivity is big, but neuronal
connectivity is very low)
Low spiking rate (0.1 Hz-1 Hz on average)[14,15]
14 P. Lennie. “The cost of cortical computation”. In: Current biology 13.6 (2003), pp. 493–497.
15 S. Herculano-Houzel. “Scaling of brain metabolism with a fixed energy budget per neuron:
implications for neuronal activity, plasticity and evolution”. In: PloS one 6.3 (2011), e17514.
18/39
Using context
Neuroscience:
Sparse graph (synaptic connectivity is big, but neuronal
connectivity is very low)
Low spiking rate (0.1 Hz-1 Hz on average)[14,15]
Large to very large networks (𝑀 >> 106 )[16]
14 P. Lennie. “The cost of cortical computation”. In: Current biology 13.6 (2003), pp. 493–497.
15 S. Herculano-Houzel. “Scaling of brain metabolism with a fixed energy budget per neuron:
implications for neuronal activity, plasticity and evolution”. In: PloS one 6.3 (2011), e17514.
16 S. Herculano-Houzel, B. Mota, and R. Lent. “Cellular scaling rules for rodent brains”. In: Proceedings
of the National Academy of Sciences 103.32 (2006), pp. 12138–12143. issn: 0027-8424. eprint:
https://www.pnas.org/content/103/32/12138.full.pdf.
18/39
Using context
Neuroscience:
Sparse graph (synaptic connectivity is big, but neuronal
connectivity is very low)
Low spiking rate (0.1 Hz-1 Hz on average)[14,15]
Large to very large networks (𝑀 >> 106 )[16]
14 P. Lennie. “The cost of cortical computation”. In: Current biology 13.6 (2003), pp. 493–497.
15 S. Herculano-Houzel. “Scaling of brain metabolism with a fixed energy budget per neuron:
implications for neuronal activity, plasticity and evolution”. In: PloS one 6.3 (2011), e17514.
16 S. Herculano-Houzel, B. Mota, and R. Lent. “Cellular scaling rules for rodent brains”. In: Proceedings
of the National Academy of Sciences 103.32 (2006), pp. 12138–12143. issn: 0027-8424. eprint:
https://www.pnas.org/content/103/32/12138.full.pdf.
18/39
Local Independence
17 V. Didelez. “Graphical models for marked point processes based on local independence”. In: Journal
of the Royal Statistical Society: Series B (Statistical Methodology) 70.1 (2008), pp. 245–264.
18 F. Figueiredo, G. Resende Borges, P. O.S. Vaz de Melo, and R. Assunção. “Fast Estimation of Causal
Interactions using Wold Processes”. In: Advances in Neural Information Processing Systems. Ed. by
S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett. Vol. 31. Curran
Associates, Inc., 2018.
19/39
Local Independence
𝑃1 𝑃2 𝑃3
17 V. Didelez. “Graphical models for marked point processes based on local independence”. In: Journal
of the Royal Statistical Society: Series B (Statistical Methodology) 70.1 (2008), pp. 245–264.
19/39
Local Independence
𝑃1 𝑃2 𝑃3
Property
The local independence graph and the interaction graph are the same
for the Hawkes process.
17 V. Didelez. “Graphical models for marked point processes based on local independence”. In: Journal
of the Royal Statistical Society: Series B (Statistical Methodology) 70.1 (2008), pp. 245–264.
19/39
Event-based simulation
21/39
Scheduler
21/39
Scheduler
21/39
Scheduler
21/39
Scheduler
21/39
Scheduler
(7, 𝑣2 )
(3, 𝑣3 ) (12, 𝑣4 )
ℎ ≤ log2 𝑁
(1, 𝑣6 ) (5, 𝑣7 ) (9, 𝑣1 ) (13, 𝑣0 )
(10, 𝑣5 )
21/39
Scheduler
21/39
How to represent functions?
22/39
How to represent functions?
22/39
How to represent functions?
𝑇1
𝑡
𝜙𝑇1 (𝑥𝑇1 )
Rejected, P ∼ Γ𝑇1
22/39
Efficient representation of a piecewise constant function
𝜙𝑡 (𝑥𝑡 )
3
𝑡
0 1 2 3 4 5
23/39
Efficient representation of a piecewise constant function
𝜙𝑡 (𝑥𝑡 )
3
𝑡
0 1 2 3 4 5
23/39
Efficient representation of a piecewise constant function
𝜙𝑡 (𝑥𝑡 ) 𝜙(𝑥𝑡 )
3 1
2
1 + 0.5
𝑡 𝑡
0 1 2 3 4 5 0 1 2 3 4 5
23/39
Efficient representation of a piecewise constant function
𝜙𝑡 (𝑥𝑡 )
4
3
2
1
𝑡
0 1 2 3 4 5
23/39
Efficient representation of a piecewise constant function
𝜙𝑡 (𝑥𝑡 )
4
3
2
1
𝑡
0 1 2 3 4 5
(0, 1) (1, 2) (1.5, 2.5) (2, 3.5) (2.5, 4) (3, 3) (3.5, 2.5) (4, 1.5) (4.5, 1) (5, 0)
23/39
Efficient representation of a piecewise constant function
𝜙𝑡 (𝑥𝑡 )
3
𝑡
0 1 2 3 4 5
23/39
Efficient representation of a piecewise constant function
𝜙𝑡 (𝑥𝑡 )
3
𝑡
0 1 2 3 4 5
23/39
Efficient representation of a piecewise constant function
𝜙𝑡 (𝑥𝑡 )
4
3
2
1
𝑡
0 1 2 3 4 5
(0, 1) (1, 1) (1.5, .5) (2, 1) (2.5, .5) (3, −1) (3.5, −.5) (4, −1) (4.5, −.5) (5, −1)
23/39
Local graph algorithm
𝑡5 𝑡3 𝑡8 𝑡1 𝑡6 …
24/39
Local graph algorithm
… 1 3
𝑡5 𝑡3 𝑡8 𝑡1 𝑡6
5 6
24/39
Local graph algorithm
24/39
Local graph algorithm
Step 3: rescheduling
… 1 3
𝑡5 𝑡1 𝑡6 𝑡3 𝑡8
5 6
𝑡1 𝑡5 𝑡8 𝑡3 𝑡9 …
8
24/39
25/39
Comparison of theoretical complexities
ℎ ℎ ℎ ℎ ℎ ℎ ℎ ℎ ℎ
⎛ ℎ11 ℎ21 ℎ31 ℎ41 ℎ51 ℎ61 ℎ71 ℎ81 ℎ91 ⎞
⎜ ℎ12 ℎ22 ℎ32 ℎ42 ℎ52 ℎ62 ℎ72 ℎ82 ℎ92 ⎟
⎜ 13 23 33 43 53 63 73 83 93 ⎟
⎜ ℎ14 ℎ24 ℎ34 ℎ44 ℎ54 ℎ64 ℎ74 ℎ84 ℎ94 ⎟
⎜ ℎ15 ℎ25 ℎ35 ℎ45 ℎ55 ℎ65 ℎ75 ℎ85 ℎ95 ⎟
⎜ ℎ16 ℎ26 ℎ36 ℎ46 ℎ56 ℎ66 ℎ76 ℎ86 ℎ96 ⎟
⎜ ℎ17 ℎ27 ℎ37 ℎ47 ℎ57 ℎ67 ℎ77 ℎ87 ℎ97 ⎟
⎜ ℎ18 ℎ28 ℎ38 ℎ48 ℎ58 ℎ68 ℎ78 ℎ88 ℎ98 ⎟
⎝ ℎ19 ℎ29 ℎ39 ℎ49 ℎ59 ℎ69 ℎ79 ℎ89 ℎ99 ⎠
28/39
Connection matrices
010110000
⎛0 0 1 0 1 1 0 0 0⎞
⎜0 0 0 0 0 1 0 0 0⎟
⎜ ⎟
⎜0 0 0 0 1 0 1 1 0⎟
⎜ 0 0 0 0 0 1 0 1 1 ⎟
⎜0 0 0 0 0 0 0 0 1⎟
⎜ 0 0 0 0 0 0 0 1 0 ⎟
⎜0 0 0 0 0 0 0 0 1⎟
⎝0 0 0 0 0 0 0 0 0⎠
28/39
Connection matrices
010110000
⎛0 0 1 0 1 1 0 0 0⎞
⎜0 0 0 0 0 1 0 0 0⎟
⎜ ⎟
⎜0 0 0 0 1 0 1 1 0⎟
⎜ 0 0 0 0 0 1 0 1 1 ⎟
⎜0 0 0 0 0 0 0 0 1⎟
⎜ 0 0 0 0 0 0 0 1 0 ⎟
⎜0 0 0 0 0 0 0 0 1⎟
⎝0 0 0 0 0 0 0 0 0⎠
28/39
Connection matrices
010110000
⎛0 0 1 0 1 1 0 0 0⎞
⎜0 0 0 0 0 1 0 0 0⎟
⎜ ⎟
⎜0 0 0 0 1 0 1 1 0⎟
⎜ 0 0 0 0 0 1 0 1 1 ⎟
⎜0 0 0 0 0 0 0 0 1⎟
⎜ 0 0 0 0 0 0 0 1 0 ⎟
⎜0 0 0 0 0 0 0 0 1⎟
⎝0 0 0 0 0 0 0 0 0⎠
Connection matrices
Full connectivity matrices can be very large
Ex: 𝑀 ∗ 𝑀 ∗ 64 =𝑀=106 64 ⋅ 1012 = 64 Tibit
28/39
Connection matrices
010110000
⎛0 0 1 0 1 1 0 0 0⎞
⎜0 0 0 0 0 1 0 0 0⎟
⎜ ⎟
⎜0 0 0 0 1 0 1 1 0⎟
⎜ 0 0 0 0 0 1 0 1 1 ⎟
⎜0 0 0 0 0 0 0 0 1⎟
⎜ 0 0 0 0 0 0 0 1 0 ⎟
⎜0 0 0 0 0 0 0 0 1⎟
⎝0 0 0 0 0 0 0 0 0⎠
Connection matrices
Full connectivity matrices can be very large
Ex: 𝑀 ∗ 𝑀 ∗ 64 =𝑀=106 64 ⋅ 1012 = 64 Tibit
But connection matrices are very sparses! Can we store only the
connections?
28/39
(Old) Yale’s format
3 245
⎛ ⎞ ⎛ ⎞
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎝ ⎠ ⎝ ⎠
Connection matrices
Full connectivity matrices can be very large
Ex: 𝑀 ∗ 𝑀 ∗ 64 =𝑀=106 64 ⋅ 1012 = 64 Tibit
But connection matrices are very sparses! Can we store only the
connections?
Yale’s format (or compressed sparse row (CSR) or compressed row
storage (CRS))
29/39
(Old) Yale’s format
3 245
⎛3⎞ ⎛3 5 6⎞
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ ⎟
⎝ ⎠ ⎝ ⎠
Connection matrices
Full connectivity matrices can be very large
Ex: 𝑀 ∗ 𝑀 ∗ 64 =𝑀=106 64 ⋅ 1012 = 64 Tibit
But connection matrices are very sparses! Can we store only the
connections?
Yale’s format (or compressed sparse row (CSR) or compressed row
storage (CRS))
29/39
(Old) Yale’s format
3 245
⎛3⎞ ⎛3 5 6⎞
⎜1⎟ ⎜6 ⎟
⎜ ⎟ ⎜ ⎟
⎜3⎟ ⎜5 7 8⎟
⎜3⎟ ⎜ 6 8 9 ⎟
⎜1⎟ ⎜9 ⎟
⎜1⎟ ⎜8 ⎟
⎜1⎟ ⎜9 ⎟
⎝0⎠ ⎝ ⎠
Connection matrices
Full connectivity matrices can be very large
Ex: 𝑀 ∗ 𝑀 ∗ 64 =𝑀=106 64 ⋅ 1012 = 64 Tibit
But connection matrices are very sparses! Can we store only the
connections?
Yale’s format (or compressed sparse row (CSR) or compressed row
storage (CRS))
29/39
(Old) Yale’s format
3 245
⎛3⎞ ⎛3 5 6⎞
⎜1⎟ ⎜6 ⎟
⎜ ⎟ ⎜ ⎟
⎜3⎟ ⎜5 7 8⎟
⎜3⎟ ⎜ 6 8 9 ⎟
⎜1⎟ ⎜9 ⎟
⎜1⎟ ⎜8 ⎟
⎜1⎟ ⎜9 ⎟
⎝0⎠ ⎝ ⎠
Connection matrices
Full connectivity matrices can be very large
Ex: 𝑀 ∗ 𝑀 ∗ 64 =𝑀=106 64 ⋅ 1012 = 64 Tibit
But connection matrices are very sparses! Can we store only the
connections?
Yale’s format (or compressed sparse row (CSR) or compressed row
storage (CRS))
Can we do better?
29/39
(Old) Yale’s format
3 245
⎛3⎞ ⎛3 5 6⎞
⎜1⎟ ⎜6 ⎟
⎜ ⎟ ⎜ ⎟
⎜3⎟ ⎜5 7 8⎟
⎜3⎟ ⎜ 6 8 9 ⎟
⎜1⎟ ⎜9 ⎟
⎜1⎟ ⎜8 ⎟
⎜1⎟ ⎜9 ⎟
⎝0⎠ ⎝ ⎠
Connection matrices
Full connectivity matrices can be very large
Ex: 𝑀 ∗ 𝑀 ∗ 64 =𝑀=106 64 ⋅ 1012 = 64 Tibit
But connection matrices are very sparses! Can we store only the
connections?
Yale’s format (or compressed sparse row (CSR) or compressed row
storage (CRS))
Can we do better?
Procedural connectivity
29/39
Procedural connectivity[22,23]
104
100
103
102 10 1
Why?
To test model fitting to experimental data
To validate simulation software
Why?
To test model fitting to experimental data
To validate simulation software
How?
Ogata’s tests (Time-rescaling theorem + ACF)
Martingale property of point processes
33/39
Ogata’s goodness-of-fit tests
33/39
Ogata’s goodness-of-fit tests
33/39
Ogata’s goodness-of-fit tests
33/39
Time transformation
34/39
Using the Martingale property of Hawkes processes
Theorem
Let N be a point process, of intensity 𝜙𝑡 (𝑥𝑡 ), with associated counting
process 𝑁𝑡 and compensator Φ𝑡 (𝑥𝑡 ). Then the difference.
𝑀𝑡 = 𝑁𝑡 − Φ𝑡 (𝑥𝑡 )
35/39
Martingale property
40
●
Difference
● ● ● ●
20 ●●● ●●● ●
●
●●
●
● ●●● ● ●●●● ● ● ●●●
●● ● ● ●●●
●● ●●● ●● ● ●
● ●●● ●● ●●●●●● ●
●●● ●● ●
●● ●●
●●●●●● ●●● ●●●● ●
● ●● ●● ●●●●●
●●
● ●●
●
●● ● ●●
●●
●●● ● ● ●●●●
●● ●
●● ●●● ●●
●●
●
●●● ● ●●●●●●● ● ●● ●● ●
●● ● ●
●●
0 ●
●●●● ●
● ●●●
●
●
●●
●●
●
●●●●● ● ●●●●●● ●●●●●●
●
●
●●●
●● ● ●●● ●●
●● ●
●●●●
●● ●
●●●● ● ●●● ●●● ●● ●●● ●
●●
● ●●
●●
●●
● ●
●●● ●●● ●● ●
● ●●
●●●●
●●●● ● ● ●
● ● ●● ●● ●● ●●
●● ●● ●●● ● ● ●● ● ● ● ●●●●
●
●
−20 ●● ● ●
● ●
40 30
20
Difference
Difference
20 10
0 0
−10
−20 −20
36/39
Conclusion
37/39
Conclusion
37/39
Conclusion
37/39
Perspectives
38/39
Perspectives
38/39
Perspectives
26 H. Mei and J. Eisner. “The neural hawkes process: A neurally self-modulating multivariate point
process”. In: arXiv preprint arXiv:1612.09328 (2016).
27 N. Du, H. Dai, R. Trivedi, U. Upadhyay, M. Gomez-Rodriguez, and L. Song. “Recurrent marked
temporal point processes: Embedding event history to vector”. In: Proceedings of the 22nd ACM
SIGKDD international conference on knowledge discovery and data mining. 2016, pp. 1555–1564.
38/39
The end?
39/39