Visible to the public Deep Location-Specific Tracking

TitleDeep Location-Specific Tracking
Publication TypeConference Paper
Year of Publication2017
AuthorsYang, Lingxiao, Liu, Risheng, Zhang, David, Zhang, Lei
Conference NameProceedings of the 25th ACM International Conference on Multimedia
PublisherACM
Conference LocationNew York, NY, USA
ISBN Number978-1-4503-4906-2
Keywordsconvolutional neural networks, deep video, location specific tracking, Metrics, pubcrawl, resilience, Resiliency, Scalability, single object tracking, visual tracking
Abstract

Convolutional Neural Network (CNN) based methods have shown significant performance gains in the problem of visual tracking in recent years. Due to many uncertain changes of objects online, such as abrupt motion, background clutter and large deformation, the visual tracking is still a challenging task. We propose a novel algorithm, namely Deep Location-Specific Tracking, which decomposes the tracking problem into a localization task and a classification task, and trains an individual network for each task. The localization network exploits the information in the current frame and provides a specific location to improve the probability of successful tracking, while the classification network finds the target among many examples generated around the target location in the previous frame, as well as the one estimated from the localization network in the current frame. CNN based trackers often have massive number of trainable parameters, and are prone to over-fitting to some particular object states, leading to less precision or tracking drift. We address this problem by learning a classification network based on 1 x 1 convolution and global average pooling. Extensive experimental results on popular benchmark datasets show that the proposed tracker achieves competitive results without using additional tracking videos for fine-tuning. The code is available at https://github.com/ZjjConan/DLST

URLhttps://dl.acm.org/doi/10.1145/3123266.3123381
DOI10.1145/3123266.3123381
Citation Keyyang_deep_2017