Schema-Tune: Noise-Driven Bias Mitigation in Transformer-Based Language Models

Abstract

In this paper, we introduce Schema-Tune, a zero-shot self-supervised framework for bias mitigation in transformer-based language models. Schema-Tune introduces curated and optimized adaptive noises to the input embeddings of transformer models to challenge the models’ embedded stereotypes. Through continuous fine-tuning steps, these noises prompt the models to change their internal semantic representations towards more socially fair representations. For fine-tuning language models, Schema-Tune relies on very limited input data: a couple of sentences formed by social group terms. Additionally, Schema-Tune defines bias and language model performance measures independently from labeled data. These measures are then used in forming the language model’s fine-tuning objective function and in searching for effective noises in the embedding space. Experimental evaluation over the StereoSet and Crows-Pairs datasets confirms that Schema-Tune is effective in mitigating bias in different social stereotype categories, including gender, race, and religion.

Cite

Text

Shokrollahi et al. "Schema-Tune: Noise-Driven Bias Mitigation in Transformer-Based Language Models." Machine Learning, 2025. doi:10.1007/S10994-024-06670-4

Markdown

[Shokrollahi et al. "Schema-Tune: Noise-Driven Bias Mitigation in Transformer-Based Language Models." Machine Learning, 2025.](https://mlanthology.org/mlj/2025/shokrollahi2025mlj-schematune/) doi:10.1007/S10994-024-06670-4

BibTeX

@article{shokrollahi2025mlj-schematune,
  title     = {{Schema-Tune: Noise-Driven Bias Mitigation in Transformer-Based Language Models}},
  author    = {Shokrollahi, Omid and Penumatcha, Ruthvik and Ensan, Faezeh and Noorian, Zeinab},
  journal   = {Machine Learning},
  year      = {2025},
  pages     = {73},
  doi       = {10.1007/S10994-024-06670-4},
  volume    = {114},
  url       = {https://mlanthology.org/mlj/2025/shokrollahi2025mlj-schematune/}
}