论文标题
特权知识蒸馏用于在线行动检测
Privileged Knowledge Distillation for Online Action Detection
论文作者
论文摘要
视频中的在线操作检测(OAD)被提议作为每个框架标签任务,以解决只能获取以前和当前视频框架的实时预测任务。本文介绍了一个新颖的学习框架,用于在线行动检测,其中仅在培训阶段可观察到的未来框架被视为特权信息的一种形式。使用知识蒸馏将特权信息从离线教师转移到在线学生中。我们注意到,此设置与常规KD不同,因为教师和学生模型之间的差异主要在于输入数据,而不是网络体系结构。我们提出了特权知识蒸馏(PKD)(i)安排课程学习程序,(ii)将辅助节点插入学生模型,既可以缩小信息差距又可以提高学习绩效。与其他明确预测未来框架的OAD方法相比,我们的方法避免学习不必要的不必要但不一致的视觉内容,并在两个流行的OAD基准,TVSeries和Thumos14上实现最先进的准确性。
Online Action Detection (OAD) in videos is proposed as a per-frame labeling task to address the real-time prediction tasks that can only obtain the previous and current video frames. This paper presents a novel learning-with-privileged based framework for online action detection where the future frames only observable at the training stages are considered as a form of privileged information. Knowledge distillation is employed to transfer the privileged information from the offline teacher to the online student. We note that this setting is different from conventional KD because the difference between the teacher and student models mostly lies in input data rather than the network architecture. We propose Privileged Knowledge Distillation (PKD) which (i) schedules a curriculum learning procedure and (ii) inserts auxiliary nodes to the student model, both for shrinking the information gap and improving learning performance. Compared to other OAD methods that explicitly predict future frames, our approach avoids learning unpredictable unnecessary yet inconsistent visual contents and achieves state-of-the-art accuracy on two popular OAD benchmarks, TVSeries and THUMOS14.