A kernel-based calculation of information on a metric space
From MaRDI portal
(Redirected from Publication:280658)
Abstract: Kernel density estimation is a technique for approximating probability distributions. Here, it is applied to the calculation of mutual information on a metric space. This is motivated by the problem in neuroscience of calculating the mutual information between stimuli and spiking responses; the space of these responses is a metric space. It is shown that kernel density estimation on a metric space resembles the k-nearest-neighbor approach. This approach is applied to a toy dataset designed to mimic electrophysiological data.
Recommendations
Cites work
- scientific article; zbMATH DE number 3789676 (Why is no real title available?)
- scientific article; zbMATH DE number 1164155 (Why is no real title available?)
- A New Multineuron Spike Train Metric
- A novel spike distance
- Estimation of Entropy and Mutual Information
- Estimation of the information by an adaptive partitioning of the observation space
- Large sample optimality of least squares cross-validation in density estimation
- On Estimation of a Probability Density Function and Mode
- Quantifying Neurotransmission Reliability Through Metrics-Based Information Analysis
- Remarks on Some Nonparametric Estimates of a Density Function
- Tight Data-Robust Bounds to Mutual Information Combining Shuffling and Model Selection Techniques
Cited in
(5)- Information-theoretic bounds and approximations in neural population coding
- Mutual information calculation using empirical classification
- The bimetric model with an informational metric tensor
- On the Choice of the Kernel Function in Kernel Discriminant Analysis Using Information Complexity
- Calculating the mutual information between two spike trains
This page was built for publication: A kernel-based calculation of information on a metric space
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q280658)