### Abstract

On-line machine learning algorithms, many biological spike-timing-dependent plasticity (STDP) learning rules, and stochastic neural dynamics evolve by Markov processes. A complete description of such systems gives the probability densities for the variables. The evolution and equilibrium state of these densities are given by a Chapman-Kolmogorov equation in discrete time, or a master equation in continuous time. These formulations are analytically intractable for most cases of interest, and to make progress a nonlinear Fokker-Planck equation (FPE) is often used in their place. The FPE is limited, and some argue that its application to describe jump processes (such as in these problems) is fundamentally flawed.We develop a well-grounded perturbation expansion that provides approximations for both the density and its moments. The approach is based on the system size expansion in statistical physics (which does . not give approximations for the density), but our simple development makes the methods accessible and invites application to diverse problems. We apply the method to calculate the equilibrium distributions for two biologically-observed STDP learning rules and for a simple nonlinear machine-learning problem. In all three examples, we show that our perturbation series provides good agreement with Monte-Carlo simulations in regimes where the FPE breaks down.

Original language | English (US) |
---|---|

Pages (from-to) | 219-228 |

Number of pages | 10 |

Journal | Neural Networks |

Volume | 32 |

DOIs | |

State | Published - Aug 1 2012 |

### Keywords

- Fokker-Planck equation
- Master equation
- Online learning
- Perturbation theory
- STDP
- State-space distributions

### ASJC Scopus subject areas

- Cognitive Neuroscience
- Artificial Intelligence

## Fingerprint Dive into the research topics of 'Approximating distributions in stochastic learning'. Together they form a unique fingerprint.

## Cite this

*Neural Networks*,

*32*, 219-228. https://doi.org/10.1016/j.neunet.2012.02.006