论文标题
受监督的句法解析对语言理解有益吗?实证研究
Is Supervised Syntactic Parsing Beneficial for Language Understanding? An Empirical Investigation
论文作者
论文摘要
长期以来,NLP长期以来一直保持(有监督的)句法解析,这是成功的高级语言理解(LU)所必需的。端到端神经模型的最新出现,通过语言建模(LM)进行了自我监督,及其在各种LU任务上的成功都提出了质疑。在这项工作中,我们从经验上研究了在LM预言变压器网络中,监督解析对语义LU的有用性。依靠已建立的微调范式,我们首先将经过验证的变压器与Biaffine解析头成为,旨在将来自通用依赖性依赖性树库的显式句法知识注入变压器中。然后,我们对LU任务的模型进行微调,并测量中间解析培训(IPT)对下游LU任务性能的影响。单语言英语和零光语言传输实验(带有中间目标解析)的结果表明,通过IPT注入变压器的显式形式化语法对下游LU性能的影响非常有限且不一致。我们的结果,再加上中间解析前后对变形金刚的表示空间的分析,在大型神经模型时代的高级语义自然语言理解中为高级语义自然语言的理解提供了重大的一步,以提供一个基本问题的答案:如何受到监督解析?
Traditional NLP has long held (supervised) syntactic parsing necessary for successful higher-level semantic language understanding (LU). The recent advent of end-to-end neural models, self-supervised via language modeling (LM), and their success on a wide range of LU tasks, however, questions this belief. In this work, we empirically investigate the usefulness of supervised parsing for semantic LU in the context of LM-pretrained transformer networks. Relying on the established fine-tuning paradigm, we first couple a pretrained transformer with a biaffine parsing head, aiming to infuse explicit syntactic knowledge from Universal Dependencies treebanks into the transformer. We then fine-tune the model for LU tasks and measure the effect of the intermediate parsing training (IPT) on downstream LU task performance. Results from both monolingual English and zero-shot language transfer experiments (with intermediate target-language parsing) show that explicit formalized syntax, injected into transformers through IPT, has very limited and inconsistent effect on downstream LU performance. Our results, coupled with our analysis of transformers' representation spaces before and after intermediate parsing, make a significant step towards providing answers to an essential question: how (un)availing is supervised parsing for high-level semantic natural language understanding in the era of large neural models?