Nonlinear Dimensionality Reduction of Data by Deep Distributed Random Samplings
Abstract
Dimensionality reduction is a fundamental problem of machine learning, and has been intensively studied, where classification and clustering are two special cases of dimensionality reduction that reduce high-dimensional data to discrete points. Here we describe a simple multilayer network for dimensionality reduction that each layer of the network is a group of mutually independent k-centers clusterings. We find that the network can be trained successfully layer-by-layer by simply assigning the centers of each clustering by randomly sampled data points from the input. Our results show that the described simple method outperformed 7 well-known dimensionality reduction methods on both very small-scale biomedical data and large-scale image and document data, with less training time than multilayer neural networks on large-scale data.
Cite
Text
Zhang. "Nonlinear Dimensionality Reduction of Data by Deep Distributed Random Samplings." Proceedings of the Sixth Asian Conference on Machine Learning, 2014.Markdown
[Zhang. "Nonlinear Dimensionality Reduction of Data by Deep Distributed Random Samplings." Proceedings of the Sixth Asian Conference on Machine Learning, 2014.](https://mlanthology.org/acml/2014/zhang2014acml-nonlinear/)BibTeX
@inproceedings{zhang2014acml-nonlinear,
title = {{Nonlinear Dimensionality Reduction of Data by Deep Distributed Random Samplings}},
author = {Zhang, Xiao-Lei},
booktitle = {Proceedings of the Sixth Asian Conference on Machine Learning},
year = {2014},
pages = {221-233},
volume = {39},
url = {https://mlanthology.org/acml/2014/zhang2014acml-nonlinear/}
}