The Information Retrieval Experiment Platform (Extended Abstract)
Abstract
With the rapid development of Large Language Models (LLMs), aligning these models with human preferences and values is critical to ensuring ethical and safe applications. However, existing alignment techniques such as RLHF or DPO often require direct fine-tuning on LLMs with billions of parameters, resulting in substantial computational costs and inefficiencies. To address this, we propose Micro token-level Accept-Reject Aligning (MARA) approach designed to operate independently of the language models. MARA simplifies the alignment process by decomposing sentence-level preference learning into token-level binary classification, where a compact three-layer fully-connected network determines whether candidate tokens are “Accepted” or “Rejected” as part of the response. Extensive experiments across seven different LLMs and three open-source datasets show that MARA achieves significant improvements in alignment performance while reducing computational costs. The source code and implementation details are publicly available at https://github.com/IAAR-Shanghai/MARA, and the trained models are released at https://huggingface.co/IAAR-Shanghai/MARA_AGENTS.
Cite
Text
Fröbe et al. "The Information Retrieval Experiment Platform (Extended Abstract)." International Joint Conference on Artificial Intelligence, 2024. doi:10.24963/ijcai.2024/931Markdown
[Fröbe et al. "The Information Retrieval Experiment Platform (Extended Abstract)." International Joint Conference on Artificial Intelligence, 2024.](https://mlanthology.org/ijcai/2024/frobe2024ijcai-information/) doi:10.24963/ijcai.2024/931BibTeX
@inproceedings{frobe2024ijcai-information,
title = {{The Information Retrieval Experiment Platform (Extended Abstract)}},
author = {Fröbe, Maik and Reimer, Jan Heinrich and MacAvaney, Sean and Deckers, Niklas and Reich, Simon and Bevendorff, Janek and Stein, Benno and Hagen, Matthias and Potthast, Martin},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2024},
pages = {8405-8410},
doi = {10.24963/ijcai.2024/931},
url = {https://mlanthology.org/ijcai/2024/frobe2024ijcai-information/}
}