DSGram: Dynamic Weighting Sub-Metrics for Grammatical Error Correction in the Era of Large Language Models

Abstract

Evaluating the performance of Grammatical Error Correction (GEC) models has become increasingly challenging, as large language model (LLM)-based GEC systems often produce corrections that diverge from provided gold references. This discrepancy undermines the reliability of traditional reference-based evaluation metrics. In this study, we propose a novel evaluation framework for GEC models, DSGram, integrating Semantic Coherence, Edit Level, and Fluency, and utilizing a dynamic weighting mechanism. Our framework employs the Analytic Hierarchy Process (AHP) in conjunction with large language models to ascertain the relative importance of various evaluation criteria. Additionally, we develop a dataset incorporating human annotations and LLM-simulated sentences to validate our algorithms and fine-tune more cost-effective models. Experimental results indicate that our proposed approach enhances the effectiveness of GEC model evaluations.

Cite

Text

Xie et al. "DSGram: Dynamic Weighting Sub-Metrics for Grammatical Error Correction in the Era of Large Language Models." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I24.34746

Markdown

[Xie et al. "DSGram: Dynamic Weighting Sub-Metrics for Grammatical Error Correction in the Era of Large Language Models." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/xie2025aaai-dsgram/) doi:10.1609/AAAI.V39I24.34746

BibTeX

@inproceedings{xie2025aaai-dsgram,
  title     = {{DSGram: Dynamic Weighting Sub-Metrics for Grammatical Error Correction in the Era of Large Language Models}},
  author    = {Xie, Jinxiang and Li, Yilin and Yin, Xunjian and Wan, Xiaojun},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {25561-25569},
  doi       = {10.1609/AAAI.V39I24.34746},
  url       = {https://mlanthology.org/aaai/2025/xie2025aaai-dsgram/}
}