University of California, San Diego
Campus Directory| Search
Daniel TartakovskyPublications › maltba-2022-autonomous
› Publications › Teaching › Research group › Useful links

Cite Details

T. E. Maltba, H. Zhao and D. M. Tartakovsky, "Autonomous learning of nonlocal stochastic neuron dynamics", J. Cogn. Neurodyn., vol. 16, doi:10.1007/s11571-021-09731-9, pp. 683-705, 2022


Neuronal dynamics is driven by externally imposed or internally generated random excitations/noise, and is often described by systems of random or stochastic ordinary differential equations. Such systems admit a distribution of solutions, which is (partially) characterized by the single-time joint probability density function (PDF) of system states. It can be used to calculate such information-theoretic quantities as the mutual information between the stochastic stimulus and various internal states of the neuron (e.g., membrane potential), as well as various spiking statistics. When random excitations are modeled as Gaussian white noise, the joint PDF of neuron states satisfies exactly a Fokker-Planck equation. However, most biologically plausible noise sources are correlated (colored). In this case, the resulting PDF equations require a closure approximation. We propose two methods for closing such equations: a modified nonlocal large-eddy-diffusivity closure and a data-driven closure relying on sparse regression to learn relevant features. The closures are tested for the stochastic non-spiking leaky integrate-and-fire and FitzHugh-Nagumo (FHN) neurons driven by sine-Wiener noise. Mutual information and total correlation between the random stimulus and the internal states of the neuron are calculated for the FHN neuron.

BibTeX Entry

author = {T. E. Maltba and H. Zhao and D. M. Tartakovsky},
title = {Autonomous learning of nonlocal stochastic neuron dynamics},
year = {2022},
urlpdf = {},
journal = {J. Cogn. Neurodyn.},
volume = {16},
doi = {10.1007/s11571-021-09731-9},
pages = {683-705}