Wednesday, October 31, 2012

Sometimes the spike-triggered average is just as good as Poisson GLMs for spike-train analysis

(hopefully no major mistakes in this one; get PDF here)

The Poisson GLM for spiking data

Generalized Linear Models (GLMs) are similar to linear regression, but account for nonlinearities and non-uniform noise in the observations. In neuroscience, it is common to predict a sequence of spikes Y={y1,..,yT}, yi{0,1}, from a series of observations X={x1,..,xT}, using a Poisson GLM:

yiPoisson(λiΔt)λi=exp(axi+b)

These models are fit by minimizing the negative log-likelihood of the observations, given the vector of regression weights a and mean offset parameter b:

argmina,bλylnλ=argmina,beaxi+by[axi+b]

Such models are usually optimized via gradient descent, or iterated re-weighted least squares. The gradient

b(...)=λya(...)=(λy)x

This is convex, and the optimum occurs at an inflection point where the gradient is zero, i.e.:

λ=yλx=yx

For simplicity, assume that x are a collection of uncorrelated zero-mean Gaussian variables. (If x has a multivariate Gaussian distribution, we an always transform into coordinates so that this is true by subtracting the mean and whitening the data.)

Linear models are often fine

What happens if we expand our likelihood at first order, and solve for a zero of the gradient? We start by expanding around the point b0=lny:

λ=eax+b=eax+bb0eb0[1+ax+bb0]eb0=[1+ax+bb0]y

Substituting this approximation in to the identity for the optimal coefficients:

[1+ax+bb0]=1[1+ax+bb0]x=yx/y

If we have chosen Δt1, so that y is indeed binary, then the right hand side of the second equation is simply the spike-triggered average. Denote this as x1=yx/y. Solving for (a,b) in the above yields:

b=b0a=x1

Which is to say: to a first approximation, the coefficients a of the Poisson GLM are simply the spike triggered average (of the mean-centered and whitened x).

Intuitively, this means if you're just trying to demonstrate some basic correlation between a variable and neuronal spiking, linear methods like the spike-triggered-average are fine. Moving to a GLM might give you more accuracy, since it has a better model of the observation process, but it often doesn't tell you much more, qualitatively.

For Gaussian covariates, GLMs capture how noise and nonlinearity interact to influence firing rate

Let's see if we can get further intuition if x is Gaussian, beyond first order. Work in normalized firing rate by defining λ¯=λ/y. The optimum can then be written as:

λ¯=1λ¯x=x1

If x is Gaussian, the expectation λ¯x has a closed form. Defining b¯=bb0=blny:

λ¯x=dxxeax+b¯(2π)D/2|Σx|1/2exp(12(xμx)Σx1(xμx)),

where D is the dimensionality of x. Working in coordinates where Σx=I and μx=0, this simplifies to:

λ¯x=(2π)D/2dxxeax+b¯12xx.

This integral can be solved by completing the square:

ax+b¯12xx=12[2ax2b¯+xx]=12[xx2ax+aaaa2b¯]=12[(xa)(xa)aa2b¯]=12[(xa)(xa)]+[12aa+b¯]λ¯x=(2π)D/2dxxe12[(xa)(xa)]+[12aa+b¯].

This is a Gaussian integral. It evaluates to:

λ¯x=ae12aa+b¯.

At the optimum, we therefore have the condition that the optimal decoding weights a should be proportional to the spike-triggered average x1:

x1=ae12aa+b¯=ae12a2+b¯ax1

Solving similarly for λ¯ gives λ¯=exp(12aa+b¯), which seems to imply that the Poisson GLM has a closed-form solution in the limit where is y sparse and x is Gaussian:

a=x1b=lny12x12

This further supports the idea that the spike-triggered average is an fast, adequate approach to statistics on spike trains. In some cases, it may be unnecessary to train a GLM.

Of course, Poisson GLMs can be fit almost as quickly as linear regressions or spike-triggered averages, using iteratively re-weighted least squares. They can also gracefully and automatically handle cases where the covariates x are not jointly Gaussian. So perhaps the question should be why not use a GLM.

No comments:

Post a Comment