Insights from a simple expression for linear fisher information in a recurrently connected population of spiking neurons.

Published

Journal Article

A simple expression for a lower bound of Fisher information is derived for a network of recurrently connected spiking neurons that have been driven to a noise-perturbed steady state. We call this lower bound linear Fisher information, as it corresponds to the Fisher information that can be recovered by a locally optimal linear estimator. Unlike recent similar calculations, the approach used here includes the effects of nonlinear gain functions and correlated input noise and yields a surprisingly simple and intuitive expression that offers substantial insight into the sources of information degradation across successive layers of a neural network. Here, this expression is used to (1) compute the optimal (i.e., information-maximizing) firing rate of a neuron, (2) demonstrate why sharpening tuning curves by either thresholding or the action of recurrent connectivity is generally a bad idea, (3) show how a single cortical expansion is sufficient to instantiate a redundant population code that can propagate across multiple cortical layers with minimal information loss, and (4) show that optimal recurrent connectivity strongly depends on the covariance structure of the inputs to the network.

Full Text

Duke Authors

Cited Authors

  • Beck, J; Bejjanki, VR; Pouget, A

Published Date

  • June 2011

Published In

Volume / Issue

  • 23 / 6

Start / End Page

  • 1484 - 1502

PubMed ID

  • 21395435

Pubmed Central ID

  • 21395435

Electronic International Standard Serial Number (EISSN)

  • 1530-888X

Digital Object Identifier (DOI)

  • 10.1162/NECO_a_00125

Language

  • eng

Conference Location

  • United States