A Sequential Learning Scheme for Function Approximation Using Minimal Radial Basis Function Neural Networks

Abstract

This article presents a sequential learning algorithm for function approximation and time-series prediction using a minimal radial basis function neural network (RBFNN). The algorithm combines the growth criterion of the resource-allocating network (RAN) of Platt (1991) with a pruning strategy based on the relative contribution of each hidden unit to the overall network output. The resulting network leads toward a minimal topology for the RBFNN. The performance of the algorithm is compared with RAN and the enhanced RAN algorithm of Kadirkamanathan and Niranjan (1993) for the following benchmark problems: (1) hearta from the benchmark problems database PROBEN1, (2) Hermite polynomial, and (3) Mackey-Glass chaotic time series. For these problems, the proposed algorithm is shown to realize RBFNNs with far fewer hidden neurons with better or same accuracy.

Cite

Text

Lu et al. "A Sequential Learning Scheme for Function Approximation Using Minimal Radial Basis Function Neural Networks." Neural Computation, 1997. doi:10.1162/NECO.1997.9.2.461

Markdown

[Lu et al. "A Sequential Learning Scheme for Function Approximation Using Minimal Radial Basis Function Neural Networks." Neural Computation, 1997.](https://mlanthology.org/neco/1997/lu1997neco-sequential/) doi:10.1162/NECO.1997.9.2.461

BibTeX

@article{lu1997neco-sequential,
  title     = {{A Sequential Learning Scheme for Function Approximation Using Minimal Radial Basis Function Neural Networks}},
  author    = {Lu, Yingwei and Sundararajan, Narasimhan and Saratchandran, Paramasivan},
  journal   = {Neural Computation},
  year      = {1997},
  pages     = {461-478},
  doi       = {10.1162/NECO.1997.9.2.461},
  volume    = {9},
  url       = {https://mlanthology.org/neco/1997/lu1997neco-sequential/}
}