论文标题

将知识从读者到猎犬提问以回答

Distilling Knowledge from Reader to Retriever for Question Answering

论文作者

Izacard, Gautier, Grave, Edouard

论文摘要

信息检索的任务是许多自然语言处理系统的重要组成部分,例如开放域问题回答。尽管传统方法是基于手工制作的特征,但基于神经网络的连续表示最近获得了竞争结果。使用此类方法的一个挑战是获取监督数据以训练回猎犬模型,与对应的查询和支持文档对应。在本文中,我们提出了一种技术,以学习以知识蒸馏的启发,并不需要带注释的查询和文档对。我们的方法利用了读者模型的注意分数,用于根据检索文档解决任务,以获取猎犬的合成标签。我们评估了我们的方法回答,获得最先进的结果。

The task of information retrieval is an important component of many natural language processing systems, such as open domain question answering. While traditional methods were based on hand-crafted features, continuous representations based on neural networks recently obtained competitive results. A challenge of using such methods is to obtain supervised data to train the retriever model, corresponding to pairs of query and support documents. In this paper, we propose a technique to learn retriever models for downstream tasks, inspired by knowledge distillation, and which does not require annotated pairs of query and documents. Our approach leverages attention scores of a reader model, used to solve the task based on retrieved documents, to obtain synthetic labels for the retriever. We evaluate our method on question answering, obtaining state-of-the-art results.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源