TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Comprehensive Study of Knowledge Editing for Large Langu...

A Comprehensive Study of Knowledge Editing for Large Language Models

Ningyu Zhang, Yunzhi Yao, Bozhong Tian, Peng Wang, Shumin Deng, Mengru Wang, Zekun Xi, Shengyu Mao, Jintian Zhang, Yuansheng Ni, Siyuan Cheng, Ziwen Xu, Xin Xu, Jia-Chen Gu, Yong Jiang, Pengjun Xie, Fei Huang, Lei Liang, Zhiqiang Zhang, Xiaowei Zhu, Jun Zhou, Huajun Chen

2024-01-02Model Editingknowledge editing
PaperPDFCode(official)Code(official)

Abstract

Large Language Models (LLMs) have shown extraordinary capabilities in understanding and generating text that closely mirrors human communication. However, a primary limitation lies in the significant computational demands during training, arising from their extensive parameterization. This challenge is further intensified by the dynamic nature of the world, necessitating frequent updates to LLMs to correct outdated information or integrate new knowledge, thereby ensuring their continued relevance. Note that many applications demand continual model adjustments post-training to address deficiencies or undesirable behaviors. There is an increasing interest in efficient, lightweight methods for on-the-fly model modifications. To this end, recent years have seen a burgeoning in the techniques of knowledge editing for LLMs, which aim to efficiently modify LLMs' behaviors within specific domains while preserving overall performance across various inputs. In this paper, we first define the knowledge editing problem and then provide a comprehensive review of cutting-edge approaches. Drawing inspiration from educational and cognitive research theories, we propose a unified categorization criterion that classifies knowledge editing methods into three groups: resorting to external knowledge, merging knowledge into the model, and editing intrinsic knowledge. Furthermore, we introduce a new benchmark, KnowEdit, for a comprehensive empirical evaluation of representative knowledge editing approaches. Additionally, we provide an in-depth analysis of knowledge location, which can give a deeper understanding of the knowledge structures inherent within LLMs. Finally, we discuss several potential applications of knowledge editing, outlining its broad and impactful implications.

Results

TaskDatasetMetricValueModel
VideozsREedit success96.74MEND
VideozsREfluency586.34MEND
VideozsRElocality92.87MEND
VideozsREportability60.41MEND
VideozsREedit success96.57ROME
VideozsRElocality27.14ROME
VideozsREportability52.2ROME
Temporal Action LocalizationzsREedit success96.74MEND
Temporal Action LocalizationzsREfluency586.34MEND
Temporal Action LocalizationzsRElocality92.87MEND
Temporal Action LocalizationzsREportability60.41MEND
Temporal Action LocalizationzsREedit success96.57ROME
Temporal Action LocalizationzsRElocality27.14ROME
Temporal Action LocalizationzsREportability52.2ROME
Zero-Shot LearningzsREedit success96.74MEND
Zero-Shot LearningzsREfluency586.34MEND
Zero-Shot LearningzsRElocality92.87MEND
Zero-Shot LearningzsREportability60.41MEND
Zero-Shot LearningzsREedit success96.57ROME
Zero-Shot LearningzsRElocality27.14ROME
Zero-Shot LearningzsREportability52.2ROME
Activity RecognitionzsREedit success96.74MEND
Activity RecognitionzsREfluency586.34MEND
Activity RecognitionzsRElocality92.87MEND
Activity RecognitionzsREportability60.41MEND
Activity RecognitionzsREedit success96.57ROME
Activity RecognitionzsRElocality27.14ROME
Activity RecognitionzsREportability52.2ROME
Action LocalizationzsREedit success96.74MEND
Action LocalizationzsREfluency586.34MEND
Action LocalizationzsRElocality92.87MEND
Action LocalizationzsREportability60.41MEND
Action LocalizationzsREedit success96.57ROME
Action LocalizationzsRElocality27.14ROME
Action LocalizationzsREportability52.2ROME
3D Action RecognitionzsREedit success96.74MEND
3D Action RecognitionzsREfluency586.34MEND
3D Action RecognitionzsRElocality92.87MEND
3D Action RecognitionzsREportability60.41MEND
3D Action RecognitionzsREedit success96.57ROME
3D Action RecognitionzsRElocality27.14ROME
3D Action RecognitionzsREportability52.2ROME
Action RecognitionzsREedit success96.74MEND
Action RecognitionzsREfluency586.34MEND
Action RecognitionzsRElocality92.87MEND
Action RecognitionzsREportability60.41MEND
Action RecognitionzsREedit success96.57ROME
Action RecognitionzsRElocality27.14ROME
Action RecognitionzsREportability52.2ROME
Model EditingzsREedit success96.74MEND
Model EditingzsREfluency586.34MEND
Model EditingzsRElocality92.87MEND
Model EditingzsREportability60.41MEND
Model EditingzsREedit success96.57ROME
Model EditingzsRElocality27.14ROME
Model EditingzsREportability52.2ROME

Related Papers

Model Editing as a Double-Edged Sword: Steering Agent Ethical Behavior Toward Beneficence or Harm2025-06-25Mitigating Safety Fallback in Editing-based Backdoor Injection on LLMs2025-06-16DualEdit: Dual Editing for Knowledge Updating in Vision-Language Models2025-06-16Model Merging for Knowledge Editing2025-06-14Mitigating Negative Interference in Multilingual Sequential Knowledge Editing through Null-Space Constraints2025-06-12SoK: Machine Unlearning for Large Language Models2025-06-10PropMEND: Hypernetworks for Knowledge Propagation in LLMs2025-06-10MEMOIR: Lifelong Model Editing with Minimal Overwrite and Informed Retention for LLMs2025-06-09