PRAE: progressive retrieval-augmented dynamic knowledge editing for large language models
Hao LI , Zheng CHU , Jiafeng LIANG , Yuxin WANG , Wei TANG , Xun MAO , Kai LV , Lei CHEN , Ming LIU , Bing QIN
Front. Comput. Sci. ›› 2027, Vol. 21 ›› Issue (1) : 2101310
The knowledge stored within large language models (LLMs) tends to become outdated as the real world rapidly evolves. Therefore, efficient knowledge editing methods have gradually been widely studied. Previous methods primarily focus on parametric knowledge injection, which is struggling to extend to large-scale editing and is time-consuming for each edit. An alternative approach is Retrieval-Augmented Generation (RAG), which enables efficient knowledge injection. However, it faces issues with conflicts between internal and external knowledge, as well as fine-grained retrieval challenges. To address this, we propose Progressive Retrieval-Augmented Dynamic Knowledge Editing (PRAE), a knowledge editing framework based on contextual knowledge injection, which fine-tunes LLMs on a carefully designed dataset to equip them with two core capabilities: progressive retrieval, enabling the step-by-step incorporation of editing knowledge to tackle multi-hop problems, and dynamic knowledge utilization, allowing the flexible and effective use of retrieved knowledge. Experimental results on seven knowledge editing datasets demonstrate that our method outperforms state-of-the-art methods by 7.1% and 25.3% on single-hop and multi-hop tasks, respectively. Our further analysis reveals that PRAE exhibits superior generalization capability and computational efficiency.
knowledge editing / large language models / retrieval-augmented generation
| [1] |
|
| [2] |
|
| [3] |
|
| [4] |
|
| [5] |
|
| [6] |
Sheng E, Chang K W, Natarajan P, Peng N. The woman worked as a babysitter: on biases in language generation. In: Proceedings of 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. 2019, 3407−3412 |
| [7] |
|
| [8] |
|
| [9] |
Yao Y, Wang P, Tian B, Cheng S, Li Z, Deng S, Chen H, Zhang N. Editing large language models: problems, methods, and opportunities. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 10222−10240 |
| [10] |
|
| [11] |
De Cao N, Aziz W, Titov I. Editing factual knowledge in language models. In: Proceedings of 2021 Conference on Empirical Methods in Natural Language Processing. 2021, 6491−6506 |
| [12] |
Mitchell E, Lin C, Bosselut A, Finn C, Manning C D. Fast model editing at scale. In: Proceedings of the 10th International Conference on Learning Representations. 2022 |
| [13] |
|
| [14] |
Meng K, Sharma A S, Andonian A J, Belinkov Y, Bau D. Mass-editing memory in a transformer. In: Proceedings of the 11th International Conference on Learning Representations. 2023 |
| [15] |
|
| [16] |
|
| [17] |
|
| [18] |
|
| [19] |
Zhong Z, Wu Z, Manning C D, Potts C, Chen D. MQuAKE: assessing knowledge editing in language models via multi-hop questions. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023. 15686−15702 |
| [20] |
Longpre S, Perisetla K, Chen A, Ramesh N, DuBois C, Singh S. Entity-based knowledge conflicts in question answering. In: Proceedings of 2021 Conference on Empirical Methods in Natural Language Processing. 2021, 7052−7063 |
| [21] |
|
| [22] |
Xie J, Zhang K, Chen J, Lou R, Su Y. Adaptive chameleon or stubborn sloth: revealing the behavior of large language models in knowledge conflicts. In: Proceedings of the 12th International Conference on Learning Representations. 2024 |
| [23] |
Chen H T, Zhang M J Q, Choi E. Rich knowledge sources bring complex knowledge conflicts: recalibrating models to reflect conflicting evidence. In: Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing. 2022, 2292−2307 |
| [24] |
Jiang Z, Xu F F, Gao L, Sun Z, Liu Q, Dwivedi-Yu J, Yang Y, Callan J, Neubig G. Active retrieval augmented generation. In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 7969−7992 |
| [25] |
|
| [26] |
|
| [27] |
Huang Z, Shen Y, Zhang X, Zhou J, Rong W, Xiong Z. Transformer-patcher: one mistake worth one neuron. In: Proceedings of the 11th International Conference on Learning Representations. 2023 |
| [28] |
Zheng C, Li L, Dong Q, Fan Y, Wu Z, Xu J, Chang B. Can we edit factual knowledge by in-context learning? In: Proceedings of 2023 Conference on Empirical Methods in Natural Language Processing. 2023, 4862−4876 |
| [29] |
Madaan A, Tandon N, Clark P, Yang Y. Memory-assisted prompt editing to improve GPT-3 after deployment. In: Proceedings of 2022 Conference on Empirical Methods in Natural Language Processing. 2022, 2833−2861 |
| [30] |
|
| [31] |
|
| [32] |
|
| [33] |
|
| [34] |
|
| [35] |
Zhou D, Schärli N, Hou L, Wei J, Scales N, Wang X, Schuurmans D, Cui C, Bousquet O, Le Q V, Chi E H. Least-to-most prompting enables complex reasoning in large language models. In: Proceedings of the 11th International Conference on Learning Representations. 2023 |
| [36] |
|
| [37] |
|
| [38] |
|
| [39] |
|
| [40] |
|
| [41] |
|
Higher Education Press
/
| 〈 |
|
〉 |