a ctive m easurement f or
play

A CTIVE M EASUREMENT F OR N EUROSCIENCE LCCC Focus Period on - PowerPoint PPT Presentation

(Towards) A CTIVE M EASUREMENT F OR N EUROSCIENCE LCCC Focus Period on Large-Scale and Distributed Optimization June 2017 Ross Boczar PhD Student boczar@berkeley.edu B erkeley C enter for C omputational Ben Eric I maging Recht


  1. (Towards) A CTIVE M EASUREMENT F OR N EUROSCIENCE 
 LCCC Focus Period on Large-Scale and Distributed Optimization June 2017 Ross Boczar PhD Student boczar@berkeley.edu … B erkeley C enter for 
 C omputational Ben Eric I maging Recht Jonas

  2. THIS TALK • Problem Statement • Motivating Science • Motivating Works • PyWren: A Shameless Plug • Some Things to Try

  3. PROBLEM STATEMENT

  4. PROBLEM STATEMENT • March 2017:

  5. PROBLEM STATEMENT • March 2017: I have to give a talk in 3 months.

  6. SCIENCE

  7. SCIENCE

  8. TWO MOTIVATORS • We have a ton of cells, and • We now have fine-grained finite experimental time! 
 control over the neurons via (can record from an organism optogenetics — we can use for a very short period of lasers to turn on and off time) individual cells or subpopulations of cells How do we learn as much about the system as quickly as possible?

  9. Start simple: SINGLE CELL 
 RESPONDING TO VISUAL INPUT

  10. MOTIVATING WORKS Sequential Optimal Experiment Design 
 for Neurophysiological Experiments 
 Lewi, Butera, and Paninski 2009 Adaptive Bayesian Methods for Closed-loop Neurophysiology 
 Pillow and Park 2016

  11. A SIMPLE EXAMPLE showing the adaptive measurement paradigm

  12. 1. present stimulus, observe response trial t ⇣ 2 σ 2 ( x − µ ) 2 ⌘ 1 f ( x ; θ ) = b + A exp − = f ( x ) λ 1 r ! λ r e � λ . p ( r | x ) = Pillow and Park 2016

  13. 2. update posterior Log-likelihood based on observed responses: L ( λ t |D t ) = log p ( R t | λ t ) = R t > log λ t − 1 > λ t , Parameter space small enough to grid in this case ( not typical ) Pillow and Park 2016

  14. 3. maximize expected utility “Infomax learning” log p ( θ | r, x , D t ) h i U infomax ( x |D t ) = E r, θ = p ( θ |D t ) One of multiple criteria to optimize (MMSE, prediction error,…) � Requires integrating over parameter and response spaces, can use MCMC / bag of samples, in this example we can numerically integrate ( not typical ) Pillow and Park 2016

  15. 1. present stimulus, observe response ... trial t+1 trial t 2. update posterior 3. maximize expected utility Pillow and Park 2016, Fig. 1

  16. In general: 1. present stimulus, observe response ... trial t+1 trial t 2. update posterior Update your belief state 3. maximize expected utility Generate the next x t in a smart, fast way Pillow and Park 2016, Fig. 1

  17. ANOTHER VIEW Lewi et al. 2009, Fig. 2

  18. ANOTHER VIEW Lewi et al. 2009, Fig. 2

  19. CURRENT APPROACH • More complicated example: Lewi-09 (visual receptive fields) • Laplace approximation for belief state (2nd order statistics) gives a compact representation for the parameter distribution

  20. CURRENT APPROACH • Have to solve high-dimensional non-convex optimization and/or integration to solve for the next x — have to grid or sample based on heuristics (i.i.d. is bad!) • Drawbacks: Curse of dimensionality, problems with EM / MCMC sampling, certain ops can get computationally (and financially!) expensive, would like to deal with more complicated models, …

  21. CURRENT APPROACH • Would like a lot of cores now , suitable for prototyping and exploration for these computationally intensive tasks, many of which are “embarrassingly parallel”

  22. PYWREN: A POSSIBLE (PARTIAL) PANACEA

  23. PREVIOUSLY, AT COMP IMAGING LUNCH Why is there no “cloud button”?

  24. My background: formerly mostly controls, now mostly ML and optimization

  25. Eric: How do you My background: get busy physicists formerly mostly and electrical controls, now mostly engineers to give ML and optimization up Matlab?

  26. PyWren pywren.io “Most wrens are small and rather inconspicuous, except for their loud and often complex songs.”

  27. PYWREN: THE API

  28. USING “SERVERLESS INFRASTRUCTURE”

  29. Powered by Continuum Analytics +

  30. 1205MB Start conda clean 977 MB eliminate pkg (Leptotyphlops carlae) 946 MB Want our runtime to include Delete non-AVX2 MKL 670 MB strip shared libs 510MB delete pyc 441MB

  31. AWS LAMBDA • 300 seconds 
 single-core (AVX2) • 512 MB in /tmp • 1.5GB RAM • Python, Java, Node

  32. AWS LAMBDA • 300 seconds 
 single-core (AVX2) • 512 MB in /tmp • 1.5GB RAM • Python, Java, Node

  33. LAMBDA SCALABILITY

  34. SOME THINGS TO TRY

  35. MPC-INSPIRED SEARCH hyperparams Current parameter distribution

  36. MPC-INSPIRED SEARCH x 1 hyperparams x 2 x 3 Current parameter distribution x 4 Possible sample locations

  37. MPC-INSPIRED SEARCH x 1 hyperparams x 2 x 3 Current parameter Dream about the future distribution x 4 Possible sample locations

  38. MPC-INSPIRED SEARCH x 1 hyperparams x 2 x 3 Current parameter Dream about the future distribution x 4 Possible sample locations n-step evaluation

  39. MPC-INSPIRED SEARCH hyperparams x 2 Current parameter distribution

  40. FUNCTION APPROXIMATION • Use Lambda services to generate rollouts to learn policies: π 1 ( x 1: t , r 1: t ; ˆ b ( θ ) t − 1 ) → ˆ b ( θ ) t Belief update function π 2 ( x 1: t , r 1: t ; ˆ b ( θ ) t ) → x t +1 Adaptive measurement function

  41. FUNCTION APPROXIMATION • Fit with ML / adaptive control / reinforcement learning / deep learning technique based on problem

  42. FUNCTION APPROXIMATION • Fit with ML / adaptive control / reinforcement learning / deep learning technique based on problem • An aside: DFO works again! 
 http://www.argmin.net/2017/04/03/evolution/

  43. THANKS! • Here all month :) • boczar@berkeley.edu • pywren.io • http://www.argmin.net/2017/04/03/evolution/

Recommend


More recommend