论文标题

无监督的常识性问题通过自我对话回答

Unsupervised Commonsense Question Answering with Self-Talk

论文作者

Shwartz, Vered, West, Peter, Bras, Ronan Le, Bhagavatula, Chandra, Choi, Yejin

论文摘要

自然语言理解涉及具有隐式背景知识的线条之间的阅读。当前的系统要么依靠预训练的语言模型作为世界知识的唯一隐式来源,要么诉诸外部知识库(KBS)来纳入其他相关知识。我们提出了一个基于自我谈话的无监督框架,作为多项选择常识任务的一种新颖替代方案。受到基于查询的发现学习的启发(Bruner,1961),我们的方法询问了语言模型,其中包含许多信息,以寻求诸如“ $ \ textit {...} $的定义是什么”,以发现其他背景知识。经验结果表明,自我对话程序显着提高了零击语言模型在六个常识基准中的四个基线的性能,并与从外部KBS获得知识的模型竞争。尽管我们的方法在几个基准上提高了性能,但即使引起正确答案的自我态度引起的知识并不总是被人类法官视为有用,从而提出了有关预识推理预训练的语言模型的内部工作的有趣问题。

Natural language understanding involves reading between the lines with implicit background knowledge. Current systems either rely on pre-trained language models as the sole implicit source of world knowledge, or resort to external knowledge bases (KBs) to incorporate additional relevant knowledge. We propose an unsupervised framework based on self-talk as a novel alternative to multiple-choice commonsense tasks. Inspired by inquiry-based discovery learning (Bruner, 1961), our approach inquires language models with a number of information seeking questions such as "$\textit{what is the definition of ...}$" to discover additional background knowledge. Empirical results demonstrate that the self-talk procedure substantially improves the performance of zero-shot language model baselines on four out of six commonsense benchmarks, and competes with models that obtain knowledge from external KBs. While our approach improves performance on several benchmarks, the self-talk induced knowledge even when leading to correct answers is not always seen as useful by human judges, raising interesting questions about the inner-workings of pre-trained language models for commonsense reasoning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源