statistical methods for understanding neural codes
play

Statistical methods for understanding neural codes Liam Paninski - PowerPoint PPT Presentation

Statistical methods for understanding neural codes Liam Paninski Department of Statistics and Center for Theoretical Neuroscience Columbia University http://www.stat.columbia.edu/ liam liam@stat.columbia.edu May 9, 2006 The neural code


  1. Statistical methods for understanding neural codes Liam Paninski Department of Statistics and Center for Theoretical Neuroscience Columbia University http://www.stat.columbia.edu/ ∼ liam liam@stat.columbia.edu May 9, 2006

  2. The neural code Input-output relationship between • External observables x (sensory stimuli, motor responses...) • Neural variables y (spike trains, population activity...) Probabilistic formulation: p ( y | x )

  3. Example: neural prosthetic design Donoghue; Cyberkinetics, Inc. ‘04 Nicolelis, Nature ’01 (Paninski et al., 1999; Serruya et al., 2002; Shoham et al., 2005)

  4. Basic goal ...learning the neural code. Fundamental question: how to estimate p ( y | x ) from experimental data? General problem is too hard — not enough data, too many inputs x and spike trains y

  5. Avoiding the curse of insufficient data Many approaches to make problem tractable: 1 : Estimate some functional f ( p ) instead e.g., information-theoretic quantities (Nemenman et al., 2002; Paninski, 2003b) 2 : Select stimuli as efficiently as possible e.g., (Foldiak, 2001; Machens, 2002; Paninski, 2003a) 3: Fit a model with small number of parameters

  6. Part 1: Neural encoding models “Encoding model”: p θ ( y | x ). — Fit parameter θ instead of full p ( y | x ) Main theme: want model to be flexible but not overly so Flexibility vs. “fittability”

  7. Multiparameter HH-type model — highly biophysically plausible, flexible — but very difficult to estimate parameters given spike times alone (figure adapted from (Fohlmeister and Miller, 1997))

  8. Cascade (“LNP”) model — easy to estimate: spike-triggered averaging (Simoncelli et al., 2004) — but not biophysically plausible (fails to capture spike timing details: refractoriness, burstiness, adaptation, etc.)

  9. Two key ideas 1. Use likelihood-based methods for fitting. — well-justified statistically — easy to incorporate prior knowledge, explicit noise models, etc. 2. Use models that are easy to fit via maximum likelihood — concave (downward-curving) functions have no non-global local maxima = ⇒ concave functions are easy to maximize by gradient ascent. Recurring theme: find flexible models whose loglikelihoods are guaranteed to be concave.

  10. Filtered integrate-and-fire model � � 0 − g ( t ) V ( t ) + I DC + � � dV ( t ) = k · � x ( t ) + h ( t − t j ) dt + σdN t ; j = −∞ (Gerstner and Kistler, 2002; Paninski et al., 2004b)

  11. Model flexibility: Adaptation

  12. The estimation problem (Paninski et al., 2004b)

  13. First passage time likelihood P (spike at t i ) = fraction of paths crossing threshold for first time at t i (computed numerically via Fokker-Planck or integral equation methods)

  14. Maximizing likelihood Maximization seems difficult, even intractable: — high-dimensional parameter space — likelihood is a complex nonlinear function of parameters Main result : The loglikelihood is concave in the parameters, no matter what data { � x ( t ) , t i } are observed. = ⇒ no non-global local maxima = ⇒ maximization easy by ascent techniques.

  15. Application: retinal ganglion cells Preparation: dissociated salamander and macaque retina — extracellularly-recorded responses of populations of RGCs Stimulus: random “flicker” visual stimuli (Chander and Chichilnisky, 2001)

  16. Spike timing precision in retina RGC LNP IF 0 0.25 0.5 0.75 1 0.07 0.17 0.22 0.26 RGC rate (sp/sec) LNP 200 IF 0 1.5 variance (sp 2 /bin) 1 0.5 0 0 0.25 0.5 0.75 1 0.6 0.64 0.85 0.9 (Pillow et al., 2005)

  17. Linking spike reliability and subthreshold noise (Pillow et al., 2005)

  18. Likelihood-based discrimination Given spike data, optimal decoder chooses stimulus � x according to likelihood: p ( spikes | � x 1 ) vs. p ( spikes | � x 2 ). Using correct model is essential (Pillow et al., 2005)

  19. Generalization: population responses Pillow et al., SFN ’05

  20. Population retinal recordings Pillow et al., SFN ’05

  21. Part 2: Decoding subthreshold activity Given extracellular spikes, what is most likely intracellular V ( t )? 10 0 −10 −20 V (mV) −30 −40 −50 −60 −70 ? −80 5.1 5.15 5.2 5.25 5.3 5.35 5.4 5.45 5.5 5.55 time (sec)

  22. Computing V ML ( t ) Loglikelihood of V ( t ) (given LIF parameters, white noise N t ): � T �� 2 L ( { V ( t ) } 0 ≤ t ≤ T ) = − 1 � � ˙ V ( t ) − − gV ( t ) + I ( t ) dt 2 σ 2 0 Constraints: • Reset at t = 0: V (0) = V reset • Spike at t = T : V ( T ) = V th • No spike for 0 < t < T : V ( t ) < V th Quadratic programming problem: optimize quadratic function under linear constraints. Concave : unique global optimum.

  23. Most likely vs. average V ( t ) 1 1 0.5 0.5 V σ =0.4 0 0 0 0.5 1 1 1 V 0.5 0.5 σ =0.2 0 0 0 0.5 1 1 1 0.5 0.5 V σ =0.1 0 0 0 0.5 1 0 0.5 1 t t (Applications to spike-triggered average (Paninski, 2005a; Paninski, 2005b))

  24. Application: in vitro data Recordings: rat sensorimotor cortical slice; dual-electrode whole-cell Stimulus: Gaussian white noise current I ( t ) Analysis: fit IF model parameters { g,� k, h ( . ) , V th , σ } by maximum likelihood (Paninski et al., 2003; Paninski et al., 2004a), then compute V ML ( t )

  25. Application: in vitro data true V(t) 0 V ML (t) −20 V (mV) −40 −60 1.04 1.05 1.06 1.07 1.08 1.09 1.1 1.11 1.12 1.13 −45 −50 −55 V (mV) −60 −65 −70 −75 1.04 1.05 1.06 1.07 1.08 1.09 1.1 1.11 1.12 1.13 time (sec) P ( V ( t ) |{ t i } , ˆ x ) computed via forward-backward hidden θ ML , � Markov model method (Paninski, 2005a).

  26. Part 3: Back to detailed models Can we recover detailed biophysical properties? • Active: membrane channel densities • Passive: axial resistances, “leakiness” of membranes • Dynamic: spatiotemporal synaptic input

  27. Conductance-based models Key point: if we observe full V i ( t ) + cell geometry, channel kinetics known + current noise is log-concave, then loglikelihood of unknown parameters is concave. Gaussian noise = ⇒ standard nonnegative regression (albeit high-d).

  28. Estimating channel densities from V ( t ) Ahrens, Huys, Paninski, NIPS ’05

  29. 0 −20 V −40 −60 50 summed currents dV/dt 0 −50 −100 20 40 60 80 100 Time [ms] conductance [mS/cm 2 ] 100 True Inferred 50 0 NaHH KHH Leak NaM KM NaS KAS Ahrens, Huys, Paninski, NIPS ’05

  30. Estimating non-homogeneous channel densities and axial resistances from spatiotemporal voltage recordings Ahrens, Huys, Paninski, COSYNE ’05

  31. Estimating synaptic inputs given V ( t ) B A Synaptic conductances Channel conductances 120 True parameters max conductance [mS/cm 2 ] 1 (spikes and conductances) 100 Data (voltage trace) Inferred (MAP) spikes Inferred (ML) channel densities 80 Inh spikes | Voltage [mV] | Exc spikes 60 0 40 20 mV 20 −25 mV 0 HHNa HHK Leak MNa MK SNa SKA SKDR C 1 −70 mV 1 0 20 mV −25 mV −70 mV 1 0 0 1280 1300 1320 1340 1360 1380 1400 0 500 1000 1500 2000 Time [ms] Time [ms] Ahrens, Huys, Paninski, NIPS ’05

  32. Collaborators Theory and numerical methods — J. Pillow, E. Simoncelli, NYU — S. Shoham, Princeton — A. Haith, C. Williams, Edinburgh — M. Ahrens, Q. Huys, Gatsby Motor cortex physiology — M. Fellows, J. Donoghue, Brown — N. Hatsopoulos, U. Chicago — B. Townsend, R. Lemon, U.C. London Retinal physiology — V. Uzzell, J. Shlens, E.J. Chichilnisky, UCSD Cortical in vitro physiology — B. Lau and A. Reyes, NYU

Recommend


More recommend