论文标题

借口对抗性学习:在自我监督的视频表示方面迈向良好实践

Pretext-Contrastive Learning: Toward Good Practices in Self-supervised Video Representation Leaning

论文作者

Tao, Li, Wang, Xueting, Yamasaki, Toshihiko

论文摘要

最近,在自我监督视频功能学习中,基于借口任务的方法是一个接一个的方法。同时,对比度学习方法也产生良好的表现。通常,新方法可以击败以前的方法,因为他们可以捕获“更好”的时间信息。但是,它们之间存在设置差异,很难得出结论。相比之下,如果这些方法尽可能接近其性能限制,将会更令人信服。在本文中,我们从一个借口任务基线开始,探索通过将其与对比度学习,数据预处理和数据扩展结合使用来走多远。从广泛的实验中发现了一个适当的设置,可以实现对基准的巨大改进,表明联合优化框架可以提高借口任务和对比度学习。我们将联合优化框架表示为借口对比度学习(PCL)。其他两个借口任务基线用于验证PCL的有效性。而且,我们可以以相同的培训方式轻松地超过当前的最新方法,以表明我们的提案的有效性和一般性。将PCL视为标准培训策略,并将其应用于自我观察的视频功能学习中的许多其他作品很方便。

Recently, pretext-task based methods are proposed one after another in self-supervised video feature learning. Meanwhile, contrastive learning methods also yield good performance. Usually, new methods can beat previous ones as claimed that they could capture "better" temporal information. However, there exist setting differences among them and it is hard to conclude which is better. It would be much more convincing in comparison if these methods have reached as closer to their performance limits as possible. In this paper, we start from one pretext-task baseline, exploring how far it can go by combining it with contrastive learning, data pre-processing, and data augmentation. A proper setting has been found from extensive experiments, with which huge improvements over the baselines can be achieved, indicating a joint optimization framework can boost both pretext task and contrastive learning. We denote the joint optimization framework as Pretext-Contrastive Learning (PCL). The other two pretext task baselines are used to validate the effectiveness of PCL. And we can easily outperform current state-of-the-art methods in the same training manner, showing the effectiveness and the generality of our proposal. It is convenient to treat PCL as a standard training strategy and apply it to many other works in self-supervised video feature learning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源