论文标题

修改变压器模型中的记忆

Modifying Memories in Transformer Models

论文作者

Zhu, Chen, Rawat, Ankit Singh, Zaheer, Manzil, Bhojanapalli, Srinadh, Li, Daliang, Yu, Felix, Kumar, Sanjiv

论文摘要

大型变压器模型在许多自然语言任务中都取得了令人印象深刻的表现。特别是,基于变压器的语言模型已被证明具有在大量参数中编码事实知识的功能。尽管已经广泛研究了变革变压器的记忆和概括的任务,但尚不众所周知,如何使变形金刚忘记特定的旧事实并记住新事实。在本文中,我们提出了\ emph {明确修改变压器模型中的特定事实知识的新任务,同时确保模型性能不会降低未修改的事实}。在许多情况下,此任务很有用,例如更新陈旧的知识,保护隐私以及消除存储在模型中的意外偏见。我们基准了几种在此任务上提供自然基线表现的方法。这导致发现了对于知识修改特别有效的变压器模型的关键组成部分。这项工作还提供了有关不同训练阶段(例如预处理和微调)对记忆和知识修改作用的作用的见解。

Large Transformer models have achieved impressive performance in many natural language tasks. In particular, Transformer based language models have been shown to have great capabilities in encoding factual knowledge in their vast amount of parameters. While the tasks of improving the memorization and generalization of Transformers have been widely studied, it is not well known how to make transformers forget specific old facts and memorize new ones. In this paper, we propose a new task of \emph{explicitly modifying specific factual knowledge in Transformer models while ensuring the model performance does not degrade on the unmodified facts}. This task is useful in many scenarios, such as updating stale knowledge, protecting privacy, and eliminating unintended biases stored in the models. We benchmarked several approaches that provide natural baseline performances on this task. This leads to the discovery of key components of a Transformer model that are especially effective for knowledge modifications. The work also provides insights into the role that different training phases (such as pretraining and fine-tuning) play towards memorization and knowledge modification.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源