论文标题
自我生成的文化学习:利用自动回归语言模型作为演示生成器
Self-Generated In-Context Learning: Leveraging Auto-regressive Language Models as a Demonstration Generator
论文作者
论文摘要
大规模的预训练语言模型(PLMS)以能够仅通过在提示中调节一些输入标签对示范而无需明确调整为所需的下游任务而在提示中配音的示范来解决任务。但是,这样的过程(即在文章中的学习)自然会高度依赖通常从外部数据集中选择的演示。在本文中,我们提出了自我生成的内在学习(SG-ICL),该学习生成了从PLM本身中的文化学习演示,以最大程度地减少对外部演示的依赖。我们对四个不同的文本分类任务进行实验,并显示SG-ICL的表现明显优于零拍的学习,并且通常价值约0.6个黄金训练样本。此外,与培训数据集的随机选择相比,我们的生成的演示表现出更一致的性能,方差较低。
Large-scale pre-trained language models (PLMs) are well-known for being capable of solving a task simply by conditioning a few input-label pairs dubbed demonstrations on a prompt without being explicitly tuned for the desired downstream task. Such a process (i.e., in-context learning), however, naturally leads to high reliance on the demonstrations which are usually selected from external datasets. In this paper, we propose self-generated in-context learning (SG-ICL), which generates demonstrations for in-context learning from PLM itself to minimize the reliance on the external demonstration. We conduct experiments on four different text classification tasks and show SG-ICL significantly outperforms zero-shot learning and is generally worth approximately 0.6 gold training samples. Moreover, our generated demonstrations show more consistent performance with low variance compared to randomly selected demonstrations from the training dataset.