Home
Scholarly Works
Nonlinear Bayesian filters for training recurrent...
Conference

Nonlinear Bayesian filters for training recurrent neural networks

Abstract

In this paper, we present nonlinear Bayesian filters for training recurrent neural networks with a special emphasis on a novel, more accurate, derivative-free member of the approximate Bayesian filter family called the cubature Kalman filter. We discuss the theory of Bayesian filters, which is rooted in the state-space modeling of the dynamic system in question and the linear estimation principle. For improved numerical stability and optimal performance during training period, a number of techniques of how to tune Bayesian filters is suggested. We compare the predictability of various Bayesian filter-trained recurrent neural networks using a chaotic time-series. From the empirical results, we conclude that the performance may be greatly improved by the new square-root cubature Kalman filter. © 2008 Springer Berlin Heidelberg.

Authors

Arasaratnam I; Haykin S

Volume

5317 LNAI

Pagination

pp. 12-33

Publication Date

December 8, 2008

DOI

10.1007/978-3-540-88636-52

Conference proceedings

Lecture Notes in Computer Science Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics

ISSN

0302-9743

Labels

View published work (Non-McMaster Users)

Contact the Experts team