Comparison of two learning networks for time series prediction
Conference Paper, Proceedings of 9th International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems (IEA/AIE '96), pp. 531 - 535, June, 1996
Abstract
Hierarchical mixtures of experts (HME) [JJ94] and radial basis function (RBF) networks [PG89] are two architectures that learn much faster than multilayer perceptrons. Their faster learning is due not to higher order search mechanisms, but to restricting the hypothesis space of the learner by constraining some of the layers of the network to use linear processing units. It can be conjectured that since their hypothesis space is restricted in the same manner, the approximation abilities of the two networks should be similar, even though their computational mechanisms are different. An empirical verification of this conjecture is presented, based on the task of predicting a nonlinear chaotic time series generated by an infrared laser.
BibTeX
@conference{Nikovski-1996-16284,author = {Daniel Nikovski and M. Zargham},
title = {Comparison of two learning networks for time series prediction},
booktitle = {Proceedings of 9th International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems (IEA/AIE '96)},
year = {1996},
month = {June},
pages = {531 - 535},
}
Copyright notice: This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. These works may not be reposted without the explicit permission of the copyright holder.