论文标题
关于生成变压器模型的任务级对话组成
On Task-Level Dialogue Composition of Generative Transformer Model
论文作者
论文摘要
面向任务的对话系统可帮助用户完成任务,例如预订电影票和通过对话订购食物。通过深神经网络参数化的生成模型被广泛用于此类系统中的下一个回答。系统的用户想在同一对话中完成多个任务是很自然的,但是生成模型构成多个任务的能力并没有很好地研究。在这项工作中,我们首先研究培训人类以任务为导向的对话的效果,即提高在变压器生成模型上构成多个任务的能力。为此,我们提出并探索两个解决方案:(1)创建合成多个任务对话数据,以从人类单一任务对话中培训培训,(2)强迫编码器代表使用辅助损失对单个和多个任务对话不变。我们实验的结果突出了即使是变形型模型的复杂变体学习从单个任务对话组成多个任务的困难。
Task-oriented dialogue systems help users accomplish tasks such as booking a movie ticket and ordering food via conversation. Generative models parameterized by a deep neural network are widely used for next turn response generation in such systems. It is natural for users of the system to want to accomplish multiple tasks within the same conversation, but the ability of generative models to compose multiple tasks is not well studied. In this work, we begin by studying the effect of training human-human task-oriented dialogues towards improving the ability to compose multiple tasks on Transformer generative models. To that end, we propose and explore two solutions: (1) creating synthetic multiple task dialogue data for training from human-human single task dialogue and (2) forcing the encoder representation to be invariant to single and multiple task dialogues using an auxiliary loss. The results from our experiments highlight the difficulty of even the sophisticated variant of transformer model in learning to compose multiple tasks from single task dialogues.