Don’t Forget the Enjoin: FocalLoRA for Instruction Hierarchical Alignment in Large Language Models
Abstract
Recent studies reveal that large language models (LLMs) often struggle to resolve conflicting instructions embedded within hierarchical prompts, resulting in decreased compliance with system-level directives and compromising the reliability of safety-critical applications. While earlier approaches attempt to improve instruction hierarchy awareness through prompt engineering or embedding-level modifications, they typically lack structural modeling and either offer limited gains or require extensive fine-tuning. In this work, we introduce $\textbf{FocalLoRA}$, a parameter-efficient and structure-aware framework that strengthens hierarchical instruction adherence by selectively optimizing structurally critical attention heads, referred to as $\textit{focal heads}$, which exhibit heightened sensitivity to instruction conflicts. Experiments across multiple models and a dedicated benchmark demonstrate that FocalLoRA markedly enhances system instruction compliance with minimal tuning cost. For instance, on Llama-8B, fine-tuning only 0.0188\% of parameters yields a 35.52\% $\uparrow$ in system instruction compliance.
Cite
Text
Shi et al. "Don’t Forget the Enjoin: FocalLoRA for Instruction Hierarchical Alignment in Large Language Models." Advances in Neural Information Processing Systems, 2025.Markdown
[Shi et al. "Don’t Forget the Enjoin: FocalLoRA for Instruction Hierarchical Alignment in Large Language Models." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/shi2025neurips-dont/)BibTeX
@inproceedings{shi2025neurips-dont,
title = {{Don’t Forget the Enjoin: FocalLoRA for Instruction Hierarchical Alignment in Large Language Models}},
author = {Shi, Zitong and Wan, Guancheng and Wang, Haixin and Li, Ruoyan and Huang, Zijie and Zhao, Wanjia and Xiao, Yijia and Luo, Xiao and Yang, Carl and Sun, Yizhou and Wang, Wei},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/shi2025neurips-dont/}
}