Multiscale Dictionary Learning for Estimating Conditional Distributions

Abstract

Nonparametric estimation of the conditional distribution of a response given high-dimensional features is a challenging problem. It is important to allow not only the mean but also the variance and shape of the response density to change flexibly with features, which are massive-dimensional. We propose a multiscale dictionary learning model, which expresses the conditional response density as a convex combination of dictionary densities, with the densities used and their weights dependent on the path through a tree decomposition of the feature space. A fast graph partitioning algorithm is applied to obtain the tree decomposition, with Bayesian methods then used to adaptively prune and average over different sub-trees in a soft probabilistic manner. The algorithm scales efficiently to approximately one million features. State of the art predictive performance is demonstrated for toy examples and two neuroscience applications including up to a million features.

Cite

Text

Petralia et al. "Multiscale Dictionary Learning for Estimating Conditional Distributions." Neural Information Processing Systems, 2013.

Markdown

[Petralia et al. "Multiscale Dictionary Learning for Estimating Conditional Distributions." Neural Information Processing Systems, 2013.](https://mlanthology.org/neurips/2013/petralia2013neurips-multiscale/)

BibTeX

@inproceedings{petralia2013neurips-multiscale,
  title     = {{Multiscale Dictionary Learning for Estimating Conditional Distributions}},
  author    = {Petralia, Francesca and Vogelstein, Joshua T and Dunson, David B},
  booktitle = {Neural Information Processing Systems},
  year      = {2013},
  pages     = {1797-1805},
  url       = {https://mlanthology.org/neurips/2013/petralia2013neurips-multiscale/}
}