Statistical Analysis of Kernel-Based Least-Squares Density-Ratio Estimation

Abstract

The ratio of two probability densities can be used for solving various machine learning tasks such as covariate shift adaptation (importance sampling), outlier detection (likelihood-ratio test), feature selection (mutual information), and conditional probability estimation. Several methods of directly estimating the density ratio have recently been developed, e.g., moment matching estimation, maximum-likelihood density-ratio estimation, and least-squares density-ratio fitting. In this paper, we propose a kernelized variant of the least-squares method for density-ratio estimation, which is called kernel unconstrained least-squares importance fitting (KuLSIF). We investigate its fundamental statistical properties including a non-parametric convergence rate, an analytic-form solution, and a leave-one-out cross-validation score. We further study its relation to other kernel-based density-ratio estimators. In experiments, we numerically compare various kernel-based density-ratio estimation methods, and show that KuLSIF compares favorably with other approaches.

Cite

Text

Kanamori et al. "Statistical Analysis of Kernel-Based Least-Squares Density-Ratio Estimation." Machine Learning, 2012. doi:10.1007/S10994-011-5266-3

Markdown

[Kanamori et al. "Statistical Analysis of Kernel-Based Least-Squares Density-Ratio Estimation." Machine Learning, 2012.](https://mlanthology.org/mlj/2012/kanamori2012mlj-statistical/) doi:10.1007/S10994-011-5266-3

BibTeX

@article{kanamori2012mlj-statistical,
  title     = {{Statistical Analysis of Kernel-Based Least-Squares Density-Ratio Estimation}},
  author    = {Kanamori, Takafumi and Suzuki, Taiji and Sugiyama, Masashi},
  journal   = {Machine Learning},
  year      = {2012},
  pages     = {335-367},
  doi       = {10.1007/S10994-011-5266-3},
  volume    = {86},
  url       = {https://mlanthology.org/mlj/2012/kanamori2012mlj-statistical/}
}