论文标题
对比度学习借口的简要介绍视觉表示
Brief Introduction to Contrastive Learning Pretext Tasks for Visual Representation
论文作者
论文摘要
为了提高视觉功能表现形式的性能,从照片或视频中用于实际应用,我们通常需要大规模的人类宣传的标记数据,同时训练深层神经网络。但是,收集和注释的人类标记的数据的成本很昂贵。鉴于现实世界中有很多未标记的数据,因此可以引入自定义的伪标签,以防止此问题。自我监督的学习,特别是对比学习,是无监督的学习方法的子集,在计算机视觉,自然语言处理和其他领域中越来越流行。对比学习的目的是将相互接近的同一样品嵌入增强样品,同时推出那些没有的样本。在以下各节中,我们将在不同的学习中介绍常规配方。在下一节中,我们将讨论各种学习的定期表述。此外,我们提供了最近出版的对比学习中的一些策略,并专注于视觉表示的借口任务。
To improve performance in visual feature representation from photos or videos for practical applications, we generally require large-scale human-annotated labeled data while training deep neural networks. However, the cost of gathering and annotating human-annotated labeled data is expensive. Given that there is a lot of unlabeled data in the actual world, it is possible to introduce self-defined pseudo labels as supervisions to prevent this issue. Self-supervised learning, specifically contrastive learning, is a subset of unsupervised learning methods that has grown popular in computer vision, natural language processing, and other domains. The purpose of contrastive learning is to embed augmented samples from the same sample near to each other while pushing away those that are not. In the following sections, we will introduce the regular formulation among different learnings. In the next sections, we will discuss the regular formulation of various learnings. Furthermore, we offer some strategies from contrastive learning that have recently been published and are focused on pretext tasks for visual representation.