CollabEdit: Towards Non-Destructive Collaborative Knowledge Editing

Abstract

Recently, collaborative fine-tuning large language models (LLMs) has emerged as a new paradigm for utilizing private data from different parties in a manner that guarantees both efficiency and privacy. Meanwhile, the practical needs of the “right to be forgotten” and the frequent demands to update outdated information, have led to a burgeoning in the techniques of knowledge editing (KE) for LLMs. However, current KE methods are all designed for a single model, and directly adapting current KE methods to collaborative learning scenarios encounters severe performance decreases. In this study, we propose a non-destructive collaborative knowledge editing framework COLLABEDIT that utilizes novel model fusion strategy to preserve overall editing performance. Empirical studies on two canonical datasets demonstrate the effectiveness and superiority of our method compared with other destructive baselines.

Cite

Text

Zheng et al. "CollabEdit: Towards Non-Destructive Collaborative Knowledge Editing." ICLR 2024 Workshops: DPFM, 2024.

Markdown

[Zheng et al. "CollabEdit: Towards Non-Destructive Collaborative Knowledge Editing." ICLR 2024 Workshops: DPFM, 2024.](https://mlanthology.org/iclrw/2024/zheng2024iclrw-collabedit/)

BibTeX

@inproceedings{zheng2024iclrw-collabedit,
  title     = {{CollabEdit: Towards Non-Destructive Collaborative Knowledge Editing}},
  author    = {Zheng, Jiamu and Zhang, Jinghuai and Wang, Futing and Du, Tianyu and Lin, Tao},
  booktitle = {ICLR 2024 Workshops: DPFM},
  year      = {2024},
  url       = {https://mlanthology.org/iclrw/2024/zheng2024iclrw-collabedit/}
}