Post

Study Notes on Neuronal Dynamics

Study Notes on Neuronal Dynamics

This is a study note for Neuronal Dynamics (Gerstner et al.).

The passive membrane

The passive membrane model can be described using an RC (resistor–capacitor) circuit, where

\[I=I_R+I_C\]

and

\[I_C=\frac{dQ}{dt}=C\cdot\frac{du}{dt}\] \[I_R=\frac{1}{R}(u-u_{rest})\]

so

\[R\cdot C\cdot \frac{du}{dt}=-(u-u_{rest})+R\cdot I(t); R\cdot C=\tau\]

If we remove the battery:

\[V=u-u_{rest}\] \[\frac{dV}{dt}=\frac{d}{dt}(u-u_{rest})=\frac{du}{dt}\] \[\tau \cdot \frac{dV}{dt}=-V+R\cdot I(t)\]

The pulse input is modelled using Dirac’s delta function:

\[I(t)=q\cdot \delta(t-t_0)\]

The total charge q is the area under the I vs. t curve.

Its effect on \(u(t)\) is to cause it jump from \(u_{rest}\) to a larger value ( \(\frac{q}{C}\) ) and then decay exponentially.

The properties of Dirac’s delta function:

\[\int_{t_0-a}^{t_0+a}\delta(t-t_0)dt=1\]

and:

\[\int_{t_0-a}^{t_0+a}f(t)\delta(t-t_0)dt=f(t_0)\]

Next we want to solve for the membrane response \(u(t)\) to the input current:

For a single pulse, suppose the duration is short:

\[\Delta=t-t_0<<\tau\]

The rise component is dominated by the first terms in the Taylor expansion of the exponential function, so:

\[u(t)=u_{rest}+\frac{q}{C}e^{-\frac{t-t_0}{\tau}}\]

For arbitrary inputs, the response is as follows:

\[u(t)=u_{rest}+\int_{-\infty}^{t}\frac{1}{C}e^{-\frac{t-t^\prime}{\tau}}I(t^\prime) dt^\prime\]

But this assumes that the input before \(t_0\) has little contributions, otherwise

\[u(t)=u_{rest}+[u(t_0)-u_{rest}]e^{-\frac{t-t_0}{\tau}}+\int_{t_0}^{t}\frac{1}{C}e^{-\frac{t-t^\prime}{\tau}}I(t^\prime) dt^\prime\]

I&F model

The previous section discussed the subthreshold regime. To simulate firing, we basically need a passive membrane + threshold.

In LIF model,

\[\tau\cdot \frac{du}{dt}=-(u-u_{rest})+R\cdot I(t)\]

If firing ( \(u(t)=\vartheta\) ), then

\[u\rightarrow u_r\]

In nonlinear I&F model,

\[\tau\cdot \frac{du}{dt}=F(u)+R\cdot I(t)\]

The \(F(u)\) can be quadratic or exponential. And the intrinsic threshold depends on the input. (Although we still have a reset condition.)

And we can draw an F (frequency)-I curve or gain function of the neuron.

Seems that the EIF model fits the experimental data (layer-5 pyramidal cells; Badel et al., J Neurophysiol., 2008) best:

\[F(u)=-(u-u_{rest})+\Delta_T e^{\frac{u-\vartheta}{\Delta_T}}\]

We didn’t consider the adaptation, noise, and other biophysical details yet.

Detailed ion-current based neuron models

According to Boltzmann distribution

\[c \propto e^{-\frac{E}{kT}}\]

and because of

\[E=q\cdot u\]

we have the Nernst equation

\[\Delta u = u_{in} - u_{out} = -\frac{kT}{q}ln\frac{c_{in}}{c_{out}}\]

The \(\Delta u\) describes a hypothetical equilibrium given the concentration difference across the cell membrane. This hypothetical equilibrium potential, or reversal potential, is specific to each ion. The total equilibrium potential is a compromise among the potentials of all ions.

\[E_{Na}\approx+67 mV; E_K\approx-83 mV\]

Hodgkin-Huxley model

The total current is the sum of capacity current, Na current, K current, and leak current (other channels):

\[I(t)=I_c + I_{Na}+I_{K}+I_l\]

Refer to the the previous notes…

\[c \cdot\frac{du}{dt}=-I_{Na}-I_K-I_l+I(t)=-\frac{1}{R_{Na}}(u-E_{Na})-\frac{1}{R_{K}}(u-E_{K})-\frac{1}{R_{l}}(u-E_{l})+I(t)\]

Considering the gating variables:

\[c \cdot\frac{du}{dt}=-g_{Na}m^3h(u-E_{Na})-g_kn^4(u-E_{K})-g_l(u-E_{l})+I(t)\]

For m, n, and h:

\[\frac{dx}{dt}=-\frac{x-x_0(u)}{\tau(u)}\]

Adaptation

Check M current and persistent sodium current.

Synapses & dendrites

Synapse

Synaptic current can be modeled by:

\[-I^{syn}(t)=-g^{syn}(t)(u-E_{syn})\]

And \(g(t)\) can be modeled by exponential decay (and rise):

\[g(t)=\sum_k \bar{g}_{syn} e^{-\frac{t-t_k}{\tau}} [1-e^{-\frac{t-t_k}{\tau_{rise}}}]\Theta(t-t_k)\]

AMPAR and GABAAR have rapid temporal dynamics, while NMDAR and GABABR are slow.

\(E_{syn}\) for excitatory synapse is ~0 mV, and ~75 mV for inhibitory synapses. The same spike would have small effect at the inhibitory synapses at resting potential. If it’s at exactly reversal potential, the driving force will be 0 (shunting inhibition).

STP

In Dayan-Abbott model (2001), the amount of “ready-to-use” neurotransmitter is represented by \(P_{rel}\).

And the maximum synaptic conductance \(\bar{g}_{syn}=c P_{rel}\).

For depression:

\[\frac{d P_{rel}}{dt}=-\frac{P_{rel}-P_0}{\tau_P}-f_D P_{rel} \sum_k \delta (t-t_k)\]

For facilitation:

\[\frac{d P_{rel}}{dt}=-\frac{P_{rel}-P_0}{\tau_P}-f_F (1-P_{rel}) \sum_k \delta (t-t_k)\]

On the right side, the first term indicates that \(P_{rel}\) would be back to the steady state \(P_0\) exponentially (with a time constant \(\tau_P\)).

Cable equation

First, the current at time \(t\), location \(x\) can be written as:

\[I(t,x)=c \frac{du(t,x)}{dt}+\sum_{ion} I_{ion}(t,x)\]

Two locations between the resistor are \(x\) and \(x+dx\), and the \(R_L\) is the longitudinal resistance.

So, \(I\) can also be described as external current injection plus current arriving and leaving:

\[I(t,x)=I^{ext}(t,x)+I_L(t,x)-L(t,x+dx)= I^{ext}(t,x)+\frac{u(t,x-dx)-2u(t,x)+u(t,x+dx)}{R_L}\]

If we downscale the $dx$:

\[R_L=r_L dx, C=c dx, I_{ion}=i_{ion} dx, I^{ext}=i^{ext} dx\]

We get the cable equation (by combining the two expressions regarding \(I(t,x)\)):

\[\frac{d^2u(t,x)}{d^2x}=c r_L \frac{d u(t,x)}{dt} + r_L \sum_{ion} i_{ion} (t,x) - r_L i^{ext}(t,x)\]

Two-dimensional models and phase plane analysis

Reduce the 4D HH model to 2D

Assumption: dendrites are approximated as passive (leak only), which leads to a point neuron.

Tricks:

  • Separation of time scales. Dynamics of m is fast, so \(m(t)=m_0(u(t)))\)

Why? For coupled differential equations

\[\begin{cases} \tau_1\frac{dx}{dt}=-x+h(y) \\ \tau_2\frac{dy}{dt}=f(y)+g(x) \end{cases}\]

If \(\tau_1\ll\tau_2\), x will approach the value of h(y) exponentially, so \(x=h(y)\)

Then \(\tau_2\frac{dy}{dt}=f(y)+g(h(y))\)

Also note that “dynamics of m is fast” means comparing with the dynamics of n, h, and I(t).

  • Dynamics of h and n are similar, so \(1-h(t)=an(t)=w(t)\)

Now HH model is reduced to

\[c\frac{du}{dt}=-g_{Na}m^3_0(u)(1-w)(u-E_{Na})-g_K[\frac{w}{a}]^4(u-E_K)-g_l(n-E_l)+I(t)\] \[\frac{dw}{dt}=-\frac{w-w_0(u)}{\tau_{eff}(u)}\]

So it’s a 2D system:

\[\begin{cases} c\frac{du}{dt}=f(u(t),w(t))+I(t) \\ \frac{dw}{dt}=g(u(t),w(t)) \end{cases}\]

The notes regarding phase plane analysis are omitted, as I don’t want to plot the figures.

Further reduction to 1D

For \(\tau_u \ll\tau_w\), the flux is nearly horizontal on the w-u phase plane.

So \(\tau_w \frac{dw}{dt}\approx 0\), which means \(w\approx w_{rest}\) when not spiking.

Then \(\tau \frac{du}{dt}=F(u,w_{rest})+RI(t)=F(u)+RI(t)\), which is the nonlinear I&F model (see previous notes).

Variability of spike trains

Source of variability

  1. Intrinsic noise from ion channels.
  2. Network noise from other neurons.

In vitro single-neuron recordings using fluctuating current stimulation showed reliable spike timing (Mainen and Sejnowski, 1995), indicating that the variability largely arises from network noise.

The variability can be described by the distribution of interspike interval (ISI), which accounts for the regularity.

Definitions of firing rate/rate codes

  1. Temporal average: \(v(t)=\frac{n^{sp}}{T}\). Examine the distribution of ISI (regularity of spike train), or the Fano factor (repeatability across trials).
  2. Average across repetitions: count the spike numbers in a very short time window. Examine the peristimulus time histogram \(PSTH(t)=\frac{n(t;t+\Delta t)}{K\Delta t}\).
  3. Population averaging: population activity \(A(t)=\frac{n(t;t+\Delta t)}{N\Delta t}\).

Seems the population activity is the one used by animals.

Homogeneous Poisson model

Assumes neuron fires at a constant rate \(\rho_0\) during time \(T\).

Probability of finding a spike in \(\Delta t\): $P_F = \rho_0 \Delta t$.

Survivor function, describing ISI: \(S(t)=e^{-\rho_0 t}; S(0)=1\). This is in continuous time steps. In discrete time steps: \(S(t)=(1-\rho_0 \Delta t)^{\frac{t}{\Delta t}}\).

Inhomogeneous Poisson model

Probability of firing: $P_F = \rho(t) \Delta t$.

Survivor function (estimate the probability of survival at time \(t\) given a spike at time \(\hat{t}\)): \(S(t\vert\hat{t})=e^{-\int_{\hat{t}}^t \rho(t^\prime) dt^\prime}\) .

Interval distribution: \(P(t\vert\hat{t})=\rho(t) S(t\vert \hat{t})\).

Stochastic spike arrival

The total spike train of \(K\) presynaptic neurons \(S(t)=\sum_{k=1}^{K} \sum_f \delta (t-t_k^f)\).

The expectation at each moment of time is the rate \(\langle S(t) \rangle = K \rho_0\).

The synaptic current \(I(t)\) sums up \(K\) neurons and firing time f.

\[\mu = \langle I(t) \rangle=\langle \frac{1}{R}\sum_k w_k \sum_f \alpha (t-t_k^f)\rangle=\frac{1}{R} \sum_k w_k \int \alpha(s) \langle \sum_f \delta (t-t_k^f-s) ds\rangle = \frac{1}{R} \sum_k w_k \int \alpha(s) ds \rho_0\]
Fluctuation of potential

As \(\langle x(t) \rangle=\int f(s) ds \rho_0\),

Autocorrelation \(\langle x(t) \hat{x}(t)\rangle = \int dt^\prime \int dt^{\prime\prime} f(t-t^\prime) f(t-t^{\prime\prime}) \langle S(t^\prime)S(t^{\prime\prime})\rangle\)

Probability of spike in step \(n\) and \(k\):

For \(n \neq k\), \(P(n,k)=P_F P_F = \rho_0^2 (\Delta t)^2\); For \(n=k\), \(P(n,n)=P_F=\rho_0 \Delta t\).

Taking \(\Delta t \rightarrow 0\), autocorrelation in continuous time steps is \(\langle S(t)S(t^{\prime})\rangle=\rho_0 \delta(t-t^\prime)+\rho_0^2\). (by dividing \(\Delta t\) in the discrete autocorrelation. The diagonal contribution (\(n=k\)) becomes infinitely sharp and turns into a Dirac delta.)

Noise models

  1. Diffusive noise.
  2. Escape noise (can estimate time-dependent interval distribution)

Escape rate (stochastic intensity of point process) \(\rho(t)=f(u(t)-\theta)\)

\(f(u-\theta)\) can be \(\frac{1}{\Delta}e^{\frac{u-\theta}{\Delta}}\) or \(ReLU(\rho_0[\frac{u}{\theta}-1])\).


Survivor function (probability to survive w/o firing again up to time \(t\)) has \(\frac{d}{dt}S_I(t\vert \hat{t})=-\rho(t)S_I(t\vert \hat{t})\), where \(\hat{t}\) is the last firing time.

So \(S_I(t\vert \hat{t})=e^{-\int_\hat{t}^t \rho(t^\prime)dt^\prime}\).

And the ISI \(P(t\vert \hat{t})=\rho(t)S(t\vert \hat{t})\).


In discrete time, the probability to survive 1 time step \(S(t_{k+1} \vert t_{k})=e^{-\rho(t_k)\Delta}\),

and the probability to fire in 1 time step \(P_k^F=1-S(t_{k+1} \vert t_{k})\).


The Likelihood of a spike train \(L(t^1,...,t^N)=e^{-\int_0^T \rho(t^\prime) dt^\prime}\prod_{f=1}^N \rho(t^f)\)

Can be transformed into log-likelihood.

Rate codes vs. temporal codes

  • Rate codes
    • Population rate
  • Temporal codes
    • Time to first spike (after input)
    • phase (wrt oscillation) of spike
    • stochastic resonance

Modern phenomenological neuron models

Add adaptation to EIF

\[\tau \frac{du}{dt}=-(u-u_{rest})+\Delta e^{\frac{u-\vartheta}{\Delta}}-R\sum_k w_k +RI(t)\]

where the dynamics of \(w_k\) is \(\tau_k \frac{dw_k}{dt}=a_k(u-u_{rest})-w_k+b_k\tau_k\sum_f \delta(t-t^f)\).

The \(a_k\) is the coupling parameter. The last term means after each spike, \(w_k\) jumps by an amount \(b_k\).

So after a spike, we do this jump and reset.

Add dynamical threshold to EIF

The threshold increases after each spike:

\[\vartheta =\theta_0+\sum_f \theta_1(t-t^f)\]

Adaptive LIF & SRM

The two linear equations (\(u\) and \(w\) dynamics) can be integrated:

\[u(t)=\sum_f \eta (t-t^f)+\int_0^{\infty}ds\kappa(s)I(t-s)+u_{rest}\]

The first term in the equation above can be rewritten as

\(\int \eta(s) \sum_f \delta(t-t^f-s)ds = \int_0^\infty \eta(s)S(t-s)ds\) where \(S(t)\) is the spike train.

Linear filters are applied for input, threshold, and refractoriness.

Neuronal populations

One population means neurons in one layer of one column.

Mean-field argument

Assume a fully connected homogeneous network (all neurons and synapses are the same; each neuron receives the same external input).

\[I_i^{syn}(t)=\sum_f \sum_k w_{ik}\alpha(t-t_k^f)=\sum_f \sum_k w_{ik}\int\alpha(s)\delta(t-t_k^f-s)ds=w_0\int \alpha(s)A(t-s)ds\]

where \(w_{ik}=\frac{w_0}{N}\).

Interpretation: all neurons receive the same total input current.

Stationary mean-field and asynchronous state

At the stationary state, for a fully connected homogeneous network with constant input:

  1. Input is constant and identical for all neurons. \(I_0=w_0 \int \alpha(s)dsA_0+I_0^{ext}\)

  2. Frequency (single-neuron gain function): \(v=g(I_0)\)

  3. Single neuron rate = population rate: \(v=A_0\)

In a stationary state of asynchronous firing, solutions might be found in the \(I_0 - A_0\) plane by combining the three equations.

Random networks

In a homogeneous network with random connectivity (with a connection probability \(p\) or fixed number of inputs \(K\)), the equations above hold.

Fokker-Planck equation for stochastic integrate-and-fire neurons

I’m unsure how to apply this knowledge. But can check the Brunel network (2000) for E-I network and dynamical regimes.

Associative memory

In anti-ferromagnet, the state of elementary magnet \(S_i\) can be +1 or -1, and the interaction \(w_{ij}\) can be +1 (if \(S_i = S_j\)) or -1 (if \(S_i \neq S_j\)). The dynamics is

\[S_i(t+1)=sgn[\sum_j w_{ij}S_j(t)]\]

To store a single pattern, the interaction can be described by the states in the prototype:

\[w_{ij}=p_ip_j\]

For multiple prototypes, you just need to sum over all prototypes (or take the average).

A minimal condition is that the pattern should stay if we start from a prototype.

\[S_i(t+1)=p_i^v sgn[1+\frac{1}{N}\sum_{\mu=1,\mu\neq v}^M \sum_{j=1}^N p_i^\mu p_i^v p_j^\mu p_j^v]=p_i^v sgn[1-a_i^v]\]

\(a_i^v\) is a random walk process with a mean of 0 and a SD of \(\sqrt{\frac{M-1}{N}}\).

Stochastic Hopfield network

The interaction is defined by

\[w_{ij}=\frac{1}{N}\sum_{\mu=1}^M p_i^\mu p_j^\mu\]

And the dynamics is defined by

\[Pr \{ S_i(t+1)=+1 \vert h_i \}=g[h_i]=g[\sum_{j=1}^N w_{ij}S_j(t)]\]

Energy function

Assume symmetric interaction and deterministic asynchronous update, the energy decreases if neuron \(k\) changes.

Realistic networks

Hopfield networks can be applied to

  1. Low mean activity.
  2. Asymmetric interactions.
  3. Separation of E/I.
  4. Low probability of connections.

Continuum models

Transients

In the high noise regime,

population activity \(A(t)=F(h(t))\) (slow transient)

and the membrane potential caused by input has \(\tau \frac{d}{dt}h(t)=-h(t)+RI(t)=-h(t)+R(I^{ext}(t)+J_0qF(h(t)))=-h(t)+RI^{ext}(t)+\gamma F(h(t))\)

In the low noise regime, transients are sharp and have oscillations.

Field equation

\[\tau \frac{d}{dt}h(x,t)=-h(x,t)+RI^{ext}(x,t)+d\int w(x-x^\prime)F(h(x^\prime,t))dx\]

where \(d\) is the density of neurons, and \(x-x^\prime\) can be expressed as distance.

Solutions for the ring model:

  1. Homogeneous solution in input driven regime (weak lateral connectivity).
  2. Bump attractor regime (strong lateral connectivity).

Synaptic plasticity and learning

Bienenstock-Cooper-Munro rule

It’s a smoothed version of pre-synaptically gated Hebb’s rule.

\[\frac{d}{dt}w_{ij}=bv_i^{post}(v_i^{post}-\vartheta)v_j^{pre}\]

where \(\vartheta=f(\bar{v}_i^{post})\).

STDP

For single pairs,

traces are updated using

\[\tau_+\frac{d}{dt}z_j^+=-z_j^++\delta(t-t_j^{pre})\] \[\tau_-\frac{d}{dt}z_i^-=-z_i^-+\delta(t-t_i^{post})\]

and

\[\frac{d}{dt}w_{ij}=aw_{ij}z_j^+\delta(t-t_i^{post})-bw_{ij}z_i^-\delta(t-t_j^{pre})\]

It’s easy to adapt to multiple pairs.

Triplet STDP model

\[\frac{d}{dt}w_{ij}=A^+z_j^+z_i^{slow}\delta(t-t_i^{post})-Bz_i^-\delta(t-t_j^{pre})\]

In the first term, pre-synaptic information is the trace, and post-synaptic information is the current spike and slow spike trace.

In the second term, pre-synaptic information is the spike, and post-synaptic information is the trace.

Assume Poisson firing, this is equivalent to the BCM model:

\[\frac{d}{dt}w_{ij}=\beta A^+ (v_i^{post})^2v_j^{pre}-\alpha B v_i^{post}v_j^{pre}\]

Zenke model

The terms in the learning rule can be classified as

  1. Homosynaptic/Hebb, if it contains pre- and post- terms.
  2. Heterosynaptic plasticity, if it’s a pure post- term.
  3. Transmitter-induced, if it’s a pure pre- term.

The plasticity rule here is

\[\frac{d}{dt}w_{ij}=a_1^{pre}v_j^{pre}-a_2^{LTD}v_j^{pre}v_i^{post}+a_3^{BCM}v_j^{pre}(v_i^{post})^2-a_4^{het}(w_{ij}-z_{ij})[v_i^{post}]^4\]

The fourth-order term cause the post activity-\(\frac{dw}{dt}\) curve to bend downwards when the post activity is high, thus stabilize the system.

This post is licensed under CC BY 4.0 by the author.