论文标题
熟练:一个辩护及时的框架
ADEPT: A DEbiasing PrompT Framework
论文作者
论文摘要
几项作品证明,Finetuning是一种适用于上下文化单词嵌入的方法。同样,具有语义含义的离散提示已证明在偏见任务中有效。由于在令牌级别上的数学表示,连续提示通常会在提供预先训练的语言模型(PLM)的情况下超越离散的提示,并提供其他特定于任务的信息。尽管如此,与离散的对应物相比,通过连续提示提示,对Debias PLM做出了相对较少的努力。此外,对于大多数改变PLM原始参数的偏见方法,一个主要问题是不仅需要减少PLM的偏差,还需要确保PLM不会失去其表示能力。填充方法通常很难维持这种平衡,因为它们倾向于猛烈地消除属性单词的含义。在本文中,我们提出了Adept,这是一种使用及时调整的Debias PLM的方法,同时保持消除偏见和确保表示能力之间的微妙平衡。为了实现这一目标,我们提出了一个新的培训标准,该标准灵感来自多种多样的学习,并为其装备一个明确的辩护术语,以优化及时的调整。此外,我们对先前提出的属性训练语料库的可靠性,质量和数量进行了几项实验,以获得更清晰的特定属性原型,这表明属性的位置和相对距离与多种词的其他单词相对距离。我们评估了一些广泛认可的基准和下游任务的熟练,并发现它在维护PLM的代表能力的同时(甚至在某些情况下都可以改善(甚至可以改善))。我们进一步可视化单词在对PLM进行辩解之前和之后的相关性,并为可见效应提供了一些可能的解释。
Several works have proven that finetuning is an applicable approach for debiasing contextualized word embeddings. Similarly, discrete prompts with semantic meanings have shown to be effective in debiasing tasks. With unfixed mathematical representation at the token level, continuous prompts usually surpass discrete ones at providing a pre-trained language model (PLM) with additional task-specific information. Despite this, relatively few efforts have been made to debias PLMs by prompt tuning with continuous prompts compared to its discrete counterpart. Furthermore, for most debiasing methods that alter a PLM's original parameters, a major problem is the need to not only decrease the bias in the PLM but also to ensure that the PLM does not lose its representation ability. Finetuning methods typically have a hard time maintaining this balance, as they tend to violently remove meanings of attribute words. In this paper, we propose ADEPT, a method to debias PLMs using prompt tuning while maintaining the delicate balance between removing biases and ensuring representation ability. To achieve this, we propose a new training criterion inspired by manifold learning and equip it with an explicit debiasing term to optimize prompt tuning. In addition, we conduct several experiments with regard to the reliability, quality, and quantity of a previously proposed attribute training corpus in order to obtain a clearer prototype of a certain attribute, which indicates the attribute's position and relative distances to other words on the manifold. We evaluate ADEPT on several widely acknowledged debiasing benchmarks and downstream tasks, and find that it achieves competitive results while maintaining (and in some cases even improving) the PLM's representation ability. We further visualize words' correlation before and after debiasing a PLM, and give some possible explanations for the visible effects.