论文标题
联合实体和关系提取与集合预测网络
Joint Entity and Relation Extraction with Set Prediction Networks
论文作者
论文摘要
联合实体和关系提取任务旨在从句子中提取所有关系三元。从本质上讲,句子中包含的关系三元组是无序的。但是,以前的基于SEQ2SEQ的模型需要在训练阶段将三元组集转换为序列。为了打破这种瓶颈,我们将关节实体和关系提取视为直接设置的预测问题,以便提取模型可以摆脱预测多个三元组的负担。为了解决此集合预测问题,我们提出了具有非自动性并行解码的变压器所特有的网络。与自动回归方法以一定顺序产生三倍的方法不同,提议的网络一次镜头直接输出了最终的三元组。此外,我们还设计了一种基于固定的损失,该损失通过两部分匹配迫使独特的预测。与横向渗透损失相比,高度惩罚三重阶的小移位,提出的两分匹配损失对于任何预测的排列都是不变的。因此,它可以通过忽略三重顺序并专注于关系类型和实体来为提出的网络提供更准确的训练信号。两个基准数据集的实验表明,我们提出的模型的表现明显胜过当前的最新方法。培训代码和训练有素的模型将在http://github.com/dianbowork/spn4re上找到。
The joint entity and relation extraction task aims to extract all relational triples from a sentence. In essence, the relational triples contained in a sentence are unordered. However, previous seq2seq based models require to convert the set of triples into a sequence in the training phase. To break this bottleneck, we treat joint entity and relation extraction as a direct set prediction problem, so that the extraction model can get rid of the burden of predicting the order of multiple triples. To solve this set prediction problem, we propose networks featured by transformers with non-autoregressive parallel decoding. Unlike autoregressive approaches that generate triples one by one in a certain order, the proposed networks directly output the final set of triples in one shot. Furthermore, we also design a set-based loss that forces unique predictions via bipartite matching. Compared with cross-entropy loss that highly penalizes small shifts in triple order, the proposed bipartite matching loss is invariant to any permutation of predictions; thus, it can provide the proposed networks with a more accurate training signal by ignoring triple order and focusing on relation types and entities. Experiments on two benchmark datasets show that our proposed model significantly outperforms current state-of-the-art methods. Training code and trained models will be available at http://github.com/DianboWork/SPN4RE.