Unsupervised Deep Representation Learning for Real-Time Tracking

Springer Science and Business Media LLC - Tập 129 - Trang 400-418 - 2020
Ning Wang1, Wengang Zhou1,2, Yibing Song3, Chao Ma4, Wei Liu3, Houqiang Li1,2
1The CAS Key Laboratory of GIPAS, University of Science and Technology of China, Hefei, China
2Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, China
3Tencent AI Lab, Shenzhen, China
4The MOE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, Shanghai, China

Tóm tắt

The advancement of visual tracking has continuously been brought by deep learning models. Typically, supervised learning is employed to train these models with expensive labeled data. In order to reduce the workload of manual annotation and learn to track arbitrary objects, we propose an unsupervised learning method for visual tracking. The motivation of our unsupervised learning is that a robust tracker should be effective in bidirectional tracking. Specifically, the tracker is able to forward localize a target object in successive frames and backtrace to its initial position in the first frame. Based on such a motivation, in the training process, we measure the consistency between forward and backward trajectories to learn a robust tracker from scratch merely using unlabeled videos. We build our framework on a Siamese correlation filter network, and propose a multi-frame validation scheme and a cost-sensitive loss to facilitate unsupervised learning. Without bells and whistles, the proposed unsupervised tracker achieves the baseline accuracy of classic fully supervised trackers while achieving a real-time speed. Furthermore, our unsupervised framework exhibits a potential in leveraging more unlabeled or weakly labeled data to further improve the tracking accuracy.

Tài liệu tham khảo