论文标题
Comps:用于测试强大的属性知识及其在预训练的语言模型中的继承的概念最小对句子
COMPS: Conceptual Minimal Pair Sentences for testing Robust Property Knowledge and its Inheritance in Pre-trained Language Models
论文作者
论文摘要
人类语义认知的一个特征是它不仅能够存储和检索通过经验观察到的概念的性质,还可以促进从上级概念(动物)到其下属(狗)的属性的继承(可以呼吸),即展示财产继承。在本文中,我们提出了Comps,这是一组最小对句子的集合,该句子共同测试了预训练的语言模型(PLM),以将其归因于概念及其演示属性继承行为的能力。对22种不同PLM的分析表明,当它们在微不足道上不同时,它们可以很容易地根据财产的概念来区分概念,但是当概念根据细微的知识表示相关时,它们会相对困难。此外,我们发现PLM可以在很大程度上证明与财产继承一致的行为,但在存在分散信息的情况下失败,这会降低许多模型的性能,有时甚至低于机会。在证明简单推理方面缺乏鲁棒性,就提出了有关PLM的能力做出正确推断的重要问题,即使它们似乎拥有先决条件知识。
A characteristic feature of human semantic cognition is its ability to not only store and retrieve the properties of concepts observed through experience, but to also facilitate the inheritance of properties (can breathe) from superordinate concepts (animal) to their subordinates (dog) -- i.e. demonstrate property inheritance. In this paper, we present COMPS, a collection of minimal pair sentences that jointly tests pre-trained language models (PLMs) on their ability to attribute properties to concepts and their ability to demonstrate property inheritance behavior. Analyses of 22 different PLMs on COMPS reveal that they can easily distinguish between concepts on the basis of a property when they are trivially different, but find it relatively difficult when concepts are related on the basis of nuanced knowledge representations. Furthermore, we find that PLMs can demonstrate behavior consistent with property inheritance to a great extent, but fail in the presence of distracting information, which decreases the performance of many models, sometimes even below chance. This lack of robustness in demonstrating simple reasoning raises important questions about PLMs' capacity to make correct inferences even when they appear to possess the prerequisite knowledge.