Home
Scholarly Works
Nonlinear Bayesian Filters for Training Recurrent...
Chapter

Nonlinear Bayesian Filters for Training Recurrent Neural Networks

Abstract

In this paper, we present nonlinear Bayesian filters for training recurrent neural networks with a special emphasis on a novel, more accurate, derivative-free member of the approximate Bayesian filter family called the cubature Kalman filter. We discuss the theory of Bayesian filters, which is rooted in the state-space modeling of the dynamic system in question and the linear estimation principle. For improved numerical stability and optimal performance during training period, a number of techniques of how to tune Bayesian filters is suggested. We compare the predictability of various Bayesian filter-trained recurrent neural networks using a chaotic time-series. From the empirical results, we conclude that the performance may be greatly improved by the new square-root cubature Kalman filter.

Authors

Arasaratnam I; Haykin S

Book title

MICAI 2008: Advances in Artificial Intelligence

Series

Lecture Notes in Computer Science

Volume

5317

Pagination

pp. 12-33

Publisher

Springer Nature

Publication Date

January 1, 2008

DOI

10.1007/978-3-540-88636-5_2
View published work (Non-McMaster Users)

Contact the Experts team