Iterative Weighted Least Squares Algorithms for Neural Networks Classifiers

Abstract

This paper discusses learning algorithms of layered neural networks from the standpoint of maximum likelihood estimation. Fisher information is explicitly calculated for the network with only one neuron. It can be interpreted as a weighted covariance matrix of input vectors. A learning algorithm is presented on the basis of Fisher's scoring method. It is shown that the algorithm can be interpreted as iterations of weighted least square method. Then those results are extended to the layered network with one hidden layer. It is also shown that Fisher information is given as a weighted covariance matrix of inputs and outputs of hidden units for this network. Tow new algorithms are proposed by utilizing this information. It is experimentally shown that the algorithms converge with fewer iterations than usual BP algorithm. Especially UFS (unitwise Fisher's scoring) method reduces to the algorithm in which each unit estimates its own weights by a weighted least squares method.

Cite

Text

Kurita. "Iterative Weighted Least Squares Algorithms for Neural Networks Classifiers." International Conference on Algorithmic Learning Theory, 1992. doi:10.1007/3-540-57369-0_29

Markdown

[Kurita. "Iterative Weighted Least Squares Algorithms for Neural Networks Classifiers." International Conference on Algorithmic Learning Theory, 1992.](https://mlanthology.org/alt/1992/kurita1992alt-iterative/) doi:10.1007/3-540-57369-0_29

BibTeX

@inproceedings{kurita1992alt-iterative,
  title     = {{Iterative Weighted Least Squares Algorithms for Neural Networks Classifiers}},
  author    = {Kurita, Takio},
  booktitle = {International Conference on Algorithmic Learning Theory},
  year      = {1992},
  pages     = {77-86},
  doi       = {10.1007/3-540-57369-0_29},
  url       = {https://mlanthology.org/alt/1992/kurita1992alt-iterative/}
}