论文标题
多阶段提示知识渊博的对话生成
Multi-Stage Prompting for Knowledgeable Dialogue Generation
论文作者
论文摘要
现有的知识对话系统通常使用预审前语言模型(LM)和大规模知识库的填充版本。这些模型通常无法概括知识库之外的主题,并且需要在每次填充时保持单独的潜在大检查站。在本文中,我们旨在通过利用验证的LM中存储的固有知识以及其强大的生成能力来解决这些局限性。我们提出了一种多阶段提示的方法,以从单个验证的LM产生知识渊博的响应。我们首先提示LM根据对话上下文产生知识。然后,我们进一步提示它根据对话上下文和先前生成的知识生成响应。结果表明,在结合知识相关性和正确性时,我们的知识生成器的表现优于最新检索模型。此外,我们的多阶段促使其在响应知识和互动方面的表现分别超过了基于填充的对话模型,分别高达10%和5%。此外,我们将模型扩展到5,300亿个参数,并表明较大的LMS提高了生成正确性得分高达10%,并且响应相关性,知识差异和参与度最多可提高10%。我们的代码可在以下网址提供:https://github.com/nvidia/megatron-lm。
Existing knowledge-grounded dialogue systems typically use finetuned versions of a pretrained language model (LM) and large-scale knowledge bases. These models typically fail to generalize on topics outside of the knowledge base, and require maintaining separate potentially large checkpoints each time finetuning is needed. In this paper, we aim to address these limitations by leveraging the inherent knowledge stored in the pretrained LM as well as its powerful generation ability. We propose a multi-stage prompting approach to generate knowledgeable responses from a single pretrained LM. We first prompt the LM to generate knowledge based on the dialogue context. Then, we further prompt it to generate responses based on the dialogue context and the previously generated knowledge. Results show that our knowledge generator outperforms the state-of-the-art retrieval-based model by 5.8% when combining knowledge relevance and correctness. In addition, our multi-stage prompting outperforms the finetuning-based dialogue model in terms of response knowledgeability and engagement by up to 10% and 5%, respectively. Furthermore, we scale our model up to 530 billion parameters and show that larger LMs improve the generation correctness score by up to 10%, and response relevance, knowledgeability and engagement by up to 10%. Our code is available at: https://github.com/NVIDIA/Megatron-LM.