A Sequential Learning Scheme for Function Approximation Using Minimal Radial Basis Function Neural Networks
Abstract
This article presents a sequential learning algorithm for function approximation and time-series prediction using a minimal radial basis function neural network (RBFNN). The algorithm combines the growth criterion of the resource-allocating network (RAN) of Platt (1991) with a pruning strategy based on the relative contribution of each hidden unit to the overall network output. The resulting network leads toward a minimal topology for the RBFNN. The performance of the algorithm is compared with RAN and the enhanced RAN algorithm of Kadirkamanathan and Niranjan (1993) for the following benchmark problems: (1) hearta from the benchmark problems database PROBEN1, (2) Hermite polynomial, and (3) Mackey-Glass chaotic time series. For these problems, the proposed algorithm is shown to realize RBFNNs with far fewer hidden neurons with better or same accuracy.
Cite
Text
Lu et al. "A Sequential Learning Scheme for Function Approximation Using Minimal Radial Basis Function Neural Networks." Neural Computation, 1997. doi:10.1162/NECO.1997.9.2.461Markdown
[Lu et al. "A Sequential Learning Scheme for Function Approximation Using Minimal Radial Basis Function Neural Networks." Neural Computation, 1997.](https://mlanthology.org/neco/1997/lu1997neco-sequential/) doi:10.1162/NECO.1997.9.2.461BibTeX
@article{lu1997neco-sequential,
title = {{A Sequential Learning Scheme for Function Approximation Using Minimal Radial Basis Function Neural Networks}},
author = {Lu, Yingwei and Sundararajan, Narasimhan and Saratchandran, Paramasivan},
journal = {Neural Computation},
year = {1997},
pages = {461-478},
doi = {10.1162/NECO.1997.9.2.461},
volume = {9},
url = {https://mlanthology.org/neco/1997/lu1997neco-sequential/}
}