Schema-Tune: Noise-Driven Bias Mitigation in Transformer-Based Language Models
Abstract
In this paper, we introduce Schema-Tune, a zero-shot self-supervised framework for bias mitigation in transformer-based language models. Schema-Tune introduces curated and optimized adaptive noises to the input embeddings of transformer models to challenge the models’ embedded stereotypes. Through continuous fine-tuning steps, these noises prompt the models to change their internal semantic representations towards more socially fair representations. For fine-tuning language models, Schema-Tune relies on very limited input data: a couple of sentences formed by social group terms. Additionally, Schema-Tune defines bias and language model performance measures independently from labeled data. These measures are then used in forming the language model’s fine-tuning objective function and in searching for effective noises in the embedding space. Experimental evaluation over the StereoSet and Crows-Pairs datasets confirms that Schema-Tune is effective in mitigating bias in different social stereotype categories, including gender, race, and religion.
Cite
Text
Shokrollahi et al. "Schema-Tune: Noise-Driven Bias Mitigation in Transformer-Based Language Models." Machine Learning, 2025. doi:10.1007/S10994-024-06670-4Markdown
[Shokrollahi et al. "Schema-Tune: Noise-Driven Bias Mitigation in Transformer-Based Language Models." Machine Learning, 2025.](https://mlanthology.org/mlj/2025/shokrollahi2025mlj-schematune/) doi:10.1007/S10994-024-06670-4BibTeX
@article{shokrollahi2025mlj-schematune,
title = {{Schema-Tune: Noise-Driven Bias Mitigation in Transformer-Based Language Models}},
author = {Shokrollahi, Omid and Penumatcha, Ruthvik and Ensan, Faezeh and Noorian, Zeinab},
journal = {Machine Learning},
year = {2025},
pages = {73},
doi = {10.1007/S10994-024-06670-4},
volume = {114},
url = {https://mlanthology.org/mlj/2025/shokrollahi2025mlj-schematune/}
}