论文标题
培训语言模型总结叙事可以改善大脑对齐
Training language models to summarize narratives improves brain alignment
论文作者
论文摘要
建立对语言更深入了解的系统是自然语言处理(NLP)的核心目标之一。为了实现这一目标,最近的作品已经开始在叙述数据集上培训语言模型,这些模型需要通过整合长篇小说来提取最关键的信息。但是,这些模型是否正在学习对文本的更深入的了解,或者模型只是学习启发式即可完成任务仍然是一个悬而未决的问题。这项工作通过转向真正理解复杂语言的一种语言处理系统,进一步研究了这一点:人脑。我们表明,培训语言模型以深入叙事理解会导致更丰富的表示,这些表示可以改善与人脑活动的一致性。我们进一步发现,针对角色名称的大脑对齐方式的改进要比其他话语特征更大,这表明这些模型正在学习重要的叙事元素。综上所述,这些结果表明,这种类型的培训确实可以带来更深入的语言理解。这些发现通过揭示了脑NLP一致性背后的一些重要因素,以及NLP的理解,即对远距离环境的理解可以改善语言建模以外,这对认知神经科学产生了后果。
Building systems that achieve a deeper understanding of language is one of the central goals of natural language processing (NLP). Towards this goal, recent works have begun to train language models on narrative datasets which require extracting the most critical information by integrating across long contexts. However, it is still an open question whether these models are learning a deeper understanding of the text, or if the models are simply learning a heuristic to complete the task. This work investigates this further by turning to the one language processing system that truly understands complex language: the human brain. We show that training language models for deeper narrative understanding results in richer representations that have improved alignment to human brain activity. We further find that the improvements in brain alignment are larger for character names than for other discourse features, which indicates that these models are learning important narrative elements. Taken together, these results suggest that this type of training can indeed lead to deeper language understanding. These findings have consequences both for cognitive neuroscience by revealing some of the significant factors behind brain-NLP alignment, and for NLP by highlighting that understanding of long-range context can be improved beyond language modeling.