direkt zum Inhalt springen

direkt zum Hauptnavigationsmenü

Sie sind hier

TU Berlin

Page Content

Monte Carlo Maximum Entropy Test

On this page you can find a MATLAB and GNU Octave implementation of a
statistical test that can assess higher-order correlations of neural
population spike counts in terms of an information theoretic analysis.
The test yields reliable results even when the number of experimental
samples is small.

Download: mcmaxenttest-1.3.zip

If you use this software for publication, please cite:

A Maximum Entropy Test for Evaluating Higher-Order Correlations in Spike Counts
Citation key Onken20120
Author Onken, A. and Dragoi, V. and Obermayer, K.
Year 2012
DOI 10.1371/journal.pcbi.1002539
Journal PLoS Computational Biology
Volume 8
Abstract Evaluating the importance of higher-order correlations of neural spike counts has been notoriously hard. A large number of samples are typically required in order to estimate higher-order correlations and resulting information theoretic quantities. In typical electrophysiology data sets with many experimental conditions, however, the number of samples in each condition is rather small. Here we describe a method that allows to quantify evidence for higher-order correlations in exactly these cases. We construct a family of reference distributions: maximum entropy distributions, which are constrained only by marginals and by linear correlations as quantified by the Pearson correlation coefficient. We devise a Monte Carlo goodness-of-fit test, which tests - for a given divergence measure of interest - whether the experimental data lead to the rejection of the null hypothesis that it was generated by one of the reference distributions. Applying our test to artificial data shows that the effects of higher-order correlations on these divergence measures can be detected even when the number of samples is small. Subsequently, we apply our method to spike count data which were recorded with multielectrode arrays from the primary visual cortex of anesthetized cat during an adaptation experiment. Using mutual information as a divergence measure we find that there are spike count bin sizes at which the maximum entropy hypothesis can be rejected for a substantial number of neuronal pairs. These results demonstrate that higher-order correlations can matter when estimating information theoretic quantities in V1. They also show that our test is able to detect their presence in typical in-vivo data sets, where the number of samples is too small to estimate higher-order correlations directly.
Bibtex Type of Publication Selected:main selected:spikes selected:publications
Link to publication Download Bibtex entry

Zusatzinformationen / Extras

Quick Access:

Schnellnavigation zur Seite über Nummerneingabe