OrdShap: Feature Position Importance for Sequential Black-Box Models
Abstract
Sequential deep learning models excel in domains with temporal or sequential dependencies, but their complexity necessitates post-hoc feature attribution methods for understanding their predictions. While existing techniques quantify feature importance, they inherently assume fixed feature ordering — conflating the effects of (1) feature values and (2) their positions within input sequences. To address this gap, we introduce OrdShap, a novel attribution method that disentangles these effects by quantifying how a model's predictions change in response to permuting feature position. We establish a game-theoretic connection between OrdShap and Sanchez-Bergantiños values, providing a theoretically grounded approach to position-sensitive attribution. Empirical results from health, natural language, and synthetic datasets highlight OrdShap's effectiveness in capturing feature value and feature position attributions, and provide deeper insight into model behavior.
Cite
Text
Hill et al. "OrdShap: Feature Position Importance for Sequential Black-Box Models." Advances in Neural Information Processing Systems, 2025.Markdown
[Hill et al. "OrdShap: Feature Position Importance for Sequential Black-Box Models." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/hill2025neurips-ordshap/)BibTeX
@inproceedings{hill2025neurips-ordshap,
title = {{OrdShap: Feature Position Importance for Sequential Black-Box Models}},
author = {Hill, Davin and Hill, Brian L. and Masoomi, Aria and Nori, Vijay S and Tillman, Robert E. and Dy, Jennifer},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/hill2025neurips-ordshap/}
}