McRank: Learning to Rank Using Multiple Classification and Gradient Boosting

Abstract

We cast the ranking problem as (1) multiple classification (“Mc”) (2) multiple or- dinal classification, which lead to computationally tractable learning algorithms for relevance ranking in Web search. We consider the DCG criterion (discounted cumulative gain), a standard quality measure in information retrieval. Our ap- proach is motivated by the fact that perfect classifications result in perfect DCG scores and the DCG errors are bounded by classification errors. We propose us- ing the Expected Relevance to convert class probabilities into ranking scores. The class probabilities are learned using a gradient boosting tree algorithm. Evalua- tions on large-scale datasets show that our approach can improve LambdaRank [5] and the regressions-based ranker [6], in terms of the (normalized) DCG scores. An efficient implementation of the boosting tree algorithm is also presented.

Cite

Text

Li et al. "McRank: Learning to Rank Using Multiple Classification and Gradient Boosting." Neural Information Processing Systems, 2007.

Markdown

[Li et al. "McRank: Learning to Rank Using Multiple Classification and Gradient Boosting." Neural Information Processing Systems, 2007.](https://mlanthology.org/neurips/2007/li2007neurips-mcrank/)

BibTeX

@inproceedings{li2007neurips-mcrank,
  title     = {{McRank: Learning to Rank Using Multiple Classification and Gradient Boosting}},
  author    = {Li, Ping and Wu, Qiang and Burges, Christopher J.},
  booktitle = {Neural Information Processing Systems},
  year      = {2007},
  pages     = {897-904},
  url       = {https://mlanthology.org/neurips/2007/li2007neurips-mcrank/}
}