LLM Alignment as Retriever Optimization: An Information Retrieval Perspective
Abstract
Large Language Models (LLMs) have revolutionized artificial intelligence with capabilities in reasoning, coding, and communication, driving innovation across industries. Their true potential depends on effective alignment to ensure correct, trustworthy and ethical behavior, addressing challenges like misinformation, hallucinations, bias and misuse. While existing Reinforcement Learning (RL)-based alignment methods are notoriously complex, direct optimization approaches offer a simpler alternative. In this work, we introduce a novel direct optimization approach for LLM alignment by drawing on established Information Retrieval (IR) principles. We present a systematic framework that bridges LLM alignment and IR methodologies, mapping LLM generation and reward models to IR’s retriever-reranker paradigm. Building on this foundation, we propose LLM Alignment as Retriever Preference Optimization (LarPO), a new alignment method that enhances overall alignment quality. Extensive experiments validate LarPO’s effectiveness with 38.9 % and 13.7 % averaged improvement on AlpacaEval2 and MixEval-Hard respectively. Our work opens new avenues for advancing LLM alignment by integrating IR foundations, offering a promising direction for future research.
Cite
Text
Jin et al. "LLM Alignment as Retriever Optimization: An Information Retrieval Perspective." Proceedings of the 42nd International Conference on Machine Learning, 2025.Markdown
[Jin et al. "LLM Alignment as Retriever Optimization: An Information Retrieval Perspective." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/jin2025icml-llm/)BibTeX
@inproceedings{jin2025icml-llm,
title = {{LLM Alignment as Retriever Optimization: An Information Retrieval Perspective}},
author = {Jin, Bowen and Yoon, Jinsung and Qin, Zhen and Wang, Ziqi and Xiong, Wei and Meng, Yu and Han, Jiawei and Arik, Sercan O},
booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
year = {2025},
pages = {28213-28232},
volume = {267},
url = {https://mlanthology.org/icml/2025/jin2025icml-llm/}
}