Statistical Analysis of Kernel-Based Least-Squares Density-Ratio Estimation
Abstract
The ratio of two probability densities can be used for solving various machine learning tasks such as covariate shift adaptation (importance sampling), outlier detection (likelihood-ratio test), feature selection (mutual information), and conditional probability estimation. Several methods of directly estimating the density ratio have recently been developed, e.g., moment matching estimation, maximum-likelihood density-ratio estimation, and least-squares density-ratio fitting. In this paper, we propose a kernelized variant of the least-squares method for density-ratio estimation, which is called kernel unconstrained least-squares importance fitting (KuLSIF). We investigate its fundamental statistical properties including a non-parametric convergence rate, an analytic-form solution, and a leave-one-out cross-validation score. We further study its relation to other kernel-based density-ratio estimators. In experiments, we numerically compare various kernel-based density-ratio estimation methods, and show that KuLSIF compares favorably with other approaches.
Cite
Text
Kanamori et al. "Statistical Analysis of Kernel-Based Least-Squares Density-Ratio Estimation." Machine Learning, 2012. doi:10.1007/S10994-011-5266-3Markdown
[Kanamori et al. "Statistical Analysis of Kernel-Based Least-Squares Density-Ratio Estimation." Machine Learning, 2012.](https://mlanthology.org/mlj/2012/kanamori2012mlj-statistical/) doi:10.1007/S10994-011-5266-3BibTeX
@article{kanamori2012mlj-statistical,
title = {{Statistical Analysis of Kernel-Based Least-Squares Density-Ratio Estimation}},
author = {Kanamori, Takafumi and Suzuki, Taiji and Sugiyama, Masashi},
journal = {Machine Learning},
year = {2012},
pages = {335-367},
doi = {10.1007/S10994-011-5266-3},
volume = {86},
url = {https://mlanthology.org/mlj/2012/kanamori2012mlj-statistical/}
}