Adversarial Attacks on Online Learning to Rank with Click Feedback
Abstract
Online learning to rank (OLTR) is a sequential decision-making problem where a learning agent selects an ordered list of items and receives feedback through user clicks. Although potential attacks against OLTR algorithms may cause serious losses in real-world applications, there is limited knowledge about adversarial attacks on OLTR. This paper studies attack strategies against multiple variants of OLTR. Our first result provides an attack strategy against the UCB algorithm on classical stochastic bandits with binary feedback, which solves the key issues caused by bounded and discrete feedback that previous works cannot handle. Building on this result, we design attack algorithms against UCB-based OLTR algorithms in position-based and cascade models. Finally, we propose a general attack strategy against any algorithm under the general click model. Each attack algorithm manipulates the learning agent into choosing the target attack item $T-o(T)$ times, incurring a cumulative cost of $o(T)$. Experiments on synthetic and real data further validate the effectiveness of our proposed attack algorithms.
Cite
Text
Zuo et al. "Adversarial Attacks on Online Learning to Rank with Click Feedback." Neural Information Processing Systems, 2023.Markdown
[Zuo et al. "Adversarial Attacks on Online Learning to Rank with Click Feedback." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/zuo2023neurips-adversarial/)BibTeX
@inproceedings{zuo2023neurips-adversarial,
title = {{Adversarial Attacks on Online Learning to Rank with Click Feedback}},
author = {Zuo, Jinhang and Zhang, Zhiyao and Wang, Zhiyong and Li, Shuai and Hajiesmaili, Mohammad and Wierman, Adam},
booktitle = {Neural Information Processing Systems},
year = {2023},
url = {https://mlanthology.org/neurips/2023/zuo2023neurips-adversarial/}
}