## Michael Buice, Ph.D.

Assistant Investigator

Michael Buice is a member of the modeling, analysis, and theory team at the Allen Institute, where he explores the implications of theories of neural processing and contributes to mathematical and data analysis. Before arriving at the Allen Institute, Buice worked in the lab of Ila Fiete at the University of Texas at Austin, where he helped derive a system size expansion for the Fisher Information for sensory and working memory systems, and developed analytic expressions for the fluctuations in attractor network models of neural networks. He held a postdoctoral research position in Carson Chow's group at the Laboratory of Biological Modeling at the National Institutes of Health (NIH). There, Buice applied kinetic theory and density functional theory to oscillator models of neural networks, answering open questions regarding the stability of asynchronous firing states in networks of finite size, a dynamical phenomenon related to the information present in the network. In addition, Buice helped construct a method for deriving equivalent reduced stochastic equations for systems with "incomplete information", such as an interacting network of neurons in which only a few neurons are actually recorded. Buice earned a Ph.D. in physics from the University of Chicago working with Jack Cowan to adapt techniques from the analysis of reaction-diffusion systems in physics to the statistics of simple models of neural networks.

## Research

Research Interests

Research Interests Perception is an ill-posed problem. Many of the sights and sounds we perceive on a regular basis are ambiguous, yet we regularly identify objects consistently by sight from a variety of angles, in multiple contexts, and even when partially occluded or in poor lighting conditions. Because of this ambiguity, perception is necessarily an act of inference, as recognized by Helmholtz in the 19th century, which combines prior knowledge with data to produce estimates about characteristics of the physical world. The neural systems that govern perception must therefore encode this prior knowledge and provide a mechanism for incorporating data from low-level sensory systems. My research interests are in identifying and understanding the mechanisms and principles that the nervous system uses to perform the inferences which allow us to perceive the world. I am particularly interested in neural implementations of Bayesian inference and mechanisms by which prior knowledge is encoded as well as the implications that coding efficiency has on the structure of neural circuits. I also wish to understand how network structure relates to network activity and how that activity corresponds to the statistics of stimuli. An important component of this endeavor is understanding the characteristics of stimuli that perceptual systems evolved to efficiently interpret, how those characteristics are represented in cortex, and how (and to what extent) they can be decoded.

**Expertise**

- Theoretical & computational neuroscience
- Statistical mechanics of neural networks
- Machine learning

**Research Programs**

- Neural coding

## Selected Publications View on PUBMED

bioRxiv

June 29, 2018

de Vries S, Lecoq J, Buice MA, Groblewski PA, Ocker GK, Oliver M, Feng F, Cain N, Ledochowitsch P, Millman D, Roll K, Garrett M, Keenan T, Kuan L, Mihalas S, Olsen S, Thompson C, Wakeman W, Waters J, Williams D, Barber C, Berbesque N, Blanchard B, Bowles N, Caldejon S, Casal L, Cho A, Cross S, Dang C, Dolbeare T, Edwards M, Galbraith J, Gaudreault N, Griffin F, Hargrave P, Howard R, Huang L, Jewell S, Keller N, Knoblich U, Larkin J, Larsen R, Lau C, Lee E, Lee F, Leon A, Li L, Long F, Luviano J, Mace K, Nguyen T, Perkins J, Robertson M, Seid S, Shea-Brown E, Shi J, Sjoquist N, Slaughterbeck C, Sullivan D, Valenza R, White C, Williford A, Witten D, Zhuang J, Zeng H, Farrell C, Ng L, Bernard A, Phillips JW, R Reid C, Koch C

Visual physiology of the Layer 4 cortical circuit in silico

PLoS Computational Biology

November 12, 2018

Arkhipov A, Gouwens NW, Billeh YN, Gratiy S, Iyer R, Wei Z, Xu Z, Berg J, Buice M, Cain N, da Costa N, de Vries S, Denman D, Durand S, Feng D, Jarsky T, Lecoq J, Lee B, Li L, Mihalas S, Ocker GK, Olsen SR, Reid RC, Soler-LLavina G, Sorensen SA, Wang Q, Waters J, Scanziani M, Koch C

Predicting how and when hidden neurons skew measured synaptic interactions

PLoS Computational Biology

October 22, 2018

Brinkman BAW, Rieke F, Shea-Brown E, Buice MA

The influence of synaptic weight distribution on neuronal population dynamics

PLoS Computational Biology

October 24, 2013

Iyer R, Menon V, Buice M, Koch C, Mihalas S

Dynamic finite size effects in spiking neural networks

PLos Computational Biology

2013

Buice M, Chow C

Effective stochastic behavior in dynamical systems with incomplete information

Physical Review

November 2011

Buice M, Chow C

Systematic fluctuation expansion for neural network activity equations

Neural Computation

February 2010

Buice M, Cowan J, Chow C

Correlations, fluctuations, and stability in a finite-size network of coupled oscillators

Physical Review

September 2007

Buice M, Chow C

A kinetic theory of coupled oscillators

Physical Review Letters

February 2007

Hildebrant E, Buice M, Chow C