Online learning in Bayesian Spiking Neurons

Levin Kuhlmann, Michael Hauser-Raspe, Jonathan Manton, David B. Grayden, Jonathan Tapson, André Van Schaik

Research output: Chapter in Book/Report/Conference proceedingConference PaperResearchpeer-review

Abstract

Bayesian Spiking Neurons (BSNs) provide a probabilistic interpretation of how neurons can perform inference and learning. Learning in a single BSN can be formulated as an online maximum-likelihood expectation-maximisation (ML-EM) algorithm. This form of learning is quite slow. Here, an alternative to this learning algorithm, called Fast Learning (FL), is presented. The FL algorithm is shown to have acceptable convergence performance when compared to the ML-EM algorithm. Moreover, for our implementations the FL algorithm is approximately 25 times faster than the ML-EM algorithm. Although only approximate, the FL algorithm therefore makes learning in hierarchical BSN networks much more tractable.

Original languageEnglish
Title of host publication2012 International Joint Conference on Neural Networks, IJCNN 2012
DOIs
Publication statusPublished - 22 Aug 2012
Event2012 Annual International Joint Conference on Neural Networks, IJCNN 2012, Part of the 2012 IEEE World Congress on Computational Intelligence, WCCI 2012 - Brisbane, QLD, Australia
Duration: 10 Jun 201215 Jun 2012

Conference

Conference2012 Annual International Joint Conference on Neural Networks, IJCNN 2012, Part of the 2012 IEEE World Congress on Computational Intelligence, WCCI 2012
CountryAustralia
CityBrisbane, QLD
Period10/06/1215/06/12

Cite this