• No results found

Deep Learning Based Visual Tracking: A Review

N/A
N/A
Protected

Academic year: 2020

Share "Deep Learning Based Visual Tracking: A Review"

Copied!
5
0
0

Loading.... (view fulltext now)

Full text

(1)

2017 2nd International Conference on Software, Multimedia and Communication Engineering (SMCE 2017) ISBN: 978-1-60595-458-5

Deep Learning Based Visual Tracking: A Review

Chun-bao LI*

,

Bo YANG and Chun-hu LI

School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, Sichuan, 610054, P.R.C

*Corresponding author

Keywords: Computer vision, Visual tracking, Deep learning.

Abstract. Visual tracking, a traditional computer vision task, has been a popular research field in recent decades. As a powerful features learning method, deep learning provides a new way for the realization of visual tracking with higher accuracy and performance. Many novel trackers that based on different network models, including auto-encoder (SAE), convolutional neural network (CNN), recurrent neural networks (RNN) deep reinforcement learning (DRL) and the fusion of them were proposed by researchers in the literatures. This paper presents a comprehensive survey on deep learning based visual tracking algorithms.

Introduction

Visual tracking, also called object tracking, is the process that detects, extracts, identifies and locates the target in a sequence of images or a video. It is a fundamental computer vision task with a wide range of real-world applications, including traffic flow monitoring, medical diagnostic, visual surveillance and human-computer interaction.

Most of the existing appearance-based tracking methods have been posed as a tracking-by-detection problem. According to the model-construction mechanism, statistical modeling is classified into three categories, including generative, discriminative and hybrid generative-discriminative. One major drawback is that they rely on low-level hand-crafted features which are incapable to capture semantic information of targets, not robust to significant appearance changes, and only have limited discriminative power [1]. While much breakthrough has been made within several decades(like [1]–[4], etc), the problem can be very challenging in many practical applications due to factors such as partial occlusion, cluttered background, fast and abrupt motion, dramatic illumination changes, and large variations in viewpoint and pose [5].

Deep learning has dramatically improved the state-of-the-art in processing text, images, video, speech and many other domains such as drug discovery and genomics, since proposed in 2006. The key advantage of deep learning is the powerful capacity of feature expressions and automatic feature extractions in large data sets. With multiple non-linear layers, say a depth of 5 to 20, a system can implement extremely intricate functions of its inputs that are simultaneously sensitive to minute details—distinguishing Samoyeds from white wolves—and insensitive to large irrelevant variations such as the background, pose, lighting and surrounding objects [6].

In [5], in order to apply deep neural networks to visual tracking, Naiyan Wang and Dit-Yan Yeung have successfully taken deep learning to a new territory of challenging applications of visual tracking. Many opportunities remain open for further research. Over the past four years, a growing number of deep learning based tracking algorithms have been proposed with demonstrated success.

(2)

Deep Learning Based Trackers

In recent years, deep learning improves the performance in computer vision by a huge margin. And in image classification, object detection, object localization, scene classification, almost all the best algorithms are based on deep learning [7]. Although the application of deep learning in visual tracking is relatively late, many novel trackers were proposed by researchers with superior accuracy and performance than traditional un-deep-learning trackers over the past four years.

In this section, a few newly proposed trackers will be introduced. First, we introduced the successful attempt to take deep learning to the new territory of visual tracking. Subsequently, we analyzed literatures based on different network models including SAE, CNN, RNN, DRL and the fusion of them.

Naiyan Wang and Dit-Yan Yeung [5] proposed a robust discriminative deep learning tracker (DLT) by putting more emphasis on effective image representation learned automatically. During the offline training stage, DLT trains a stacked denoising auto-encoder (SDAE) offline to learn generic image features that are more robust against variations. The basic building block of an SDAE is a one-layer neural network. After offline training, the encoder part of the SDAE is used as a feature extractor during the online tracking process to train a classification neural network to distinguish the tracked object from the background. Finally, both the feature extractor and the classifier can be further tuned to adapt to appearance changes of the moving object. Compared with the traditional trackers, DLT has some key differences and advantages. First, DLT learns generic image features from a larger and more general dataset rather than a smaller set with only some chosen image categories. Second, DLT learns the image features from raw images automatically instead of relying on handcrafted SIFT features. Third, further learning is allowed during the online tracking process of DLT so as to adapt better to the specific object being tracked.

Naiyan Wang and his copartners [8] first presented the CNN architecture in structured output deep learning tracker (SO-DLT) and the idea that offline pre-train and online fine-tune of the CNN. Although the accuracy and performance of SO-DLT has demonstrated substantial improvement over other state-of-the-art trackers, some failure cases do exist. The tracker is likely to drift when the initial bounding box is not properly specified or when there exist distractors in the background and the target is occluded.

In order to reduce or solve the impact of the failure cases and improve the accuracy and performance of trackers, more advanced models have been investigated.

In [9], Hyeonseob Nam and Bohyung Han proposed a novel tracking algorithm based on a CNN trained in a multi-domain learning framework, which is referred to as multi-domain network (MDNet). The network mentioned above consists of shared layers and K branches of domain-specific layers. The algorithm pre-trains the MDNet using a large set of videos with tracking ground-truths to obtain a generic target representation in the shared layers. The entire network is pre-trained offline, and the fully connected layers including a single domain-specific layer are fine-tuned online. In order to avoid drifting, MDNet adopts some strategies from the detection task, such as hard negative mining, bounding-box regression etc. The proposed algorithm illustrates outstanding performance compared with state-of-the-art methods in existing tracking benchmarks.

(3)

In order to overcome challenges of the unpredictable and drastic changes in the appearance of objects, existing methods update the appearance model on-line. But inaccurate and erroneous updates of the appearance model result in a tracker drift. In [11], Janghoon Choi, Junseok Kwon and Kyoung Mu Lee introduce a novel visual tracking algorithm based on a template selection strategy constructed by deep reinforcement learning methods. The tracking algorithm utilizes this strategy to choose the best template for tracking a given frame. The template selection strategy is self-learned by utilizing a simple policy gradient method on numerous training episodes randomly generated from a tracking benchmark dataset. Although the tracking algorithm effectively decides the best template for visual tracking, the accuracy of the tracker needs to beenhanced.

[image:3.595.67.530.330.480.2]

Da Zhang, Hamid Maei, Xin Wang, and Yuan-Fang Wang presented the first neural-network tracker that combines convolutional and recurrent networks with RL algorithm in [12]. The tracker is capable of effectively leveraging temporal and contextual information among consecutive frames. And it comprises three components: a CNN extracting best tracking features in each video frame, an RNN constructing video memory state, and a reinforcement learning (RL) agent making target location decisions. The tracking problem is formulated as a decision-making process, and the model can be trained with RL algorithms to learn good tracking policies that pay attention to continuous, inter-frame correlation and maximize tracking performance in the long run. The proposed tracking approach works well in various scenarios on artificial video sequences with ground truth.

Figure 1. Precision and success plots on OTB100. The numbers in the legend indicate the representative precisions at 20 pixels for precision plots, and the area-under-curve scores for success plots.

Figure 2. Precision and success plots on OTB2015.

[image:3.595.73.517.519.700.2]
(4)

have also been analyzed in the literatures mentioned above, which facilitated a better understanding of the state-of-the-art object tracking approaches, and provided a platform for gauging new algorithms.

Conclusions

Nowadays, visual tracking, as a fundamental computer vision task, has been widely used in traffic flow monitoring, medical diagnostic, visual surveillance, human-computer interaction and robotics. Although visual tracking with deep learning starts late and develops slow, it is eye-catching in terms of the accuracy and overall performance. But it is still not so perfect, like drifting, and not real-time, etc. In this paper, many of the current important and outstanding deep learning based visual tracking algorithms have been presented and analyzed. These trackers are future extended and may provide more room for further improvement in accuracy and performance.

Acknowledgement

This work is supported by Sichuan Provincial Project of International Scientific and Technical Exchange and Research Collaboration Programs (Project No. 2016HH0023).

References

[1] X. Li, W. Hu, C. Shen, Z. Zhang, A. Dick, A.V.D. Hengel, A survey of appearance models in visual object tracking, J. ACM transactions on Intelligent Systems and Technology (TIST) 4.4 (2013): 58.

[2] C. Ma, X. Yang, C. Zhang, M. H. Yang, Long-term correlation tracking C. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. (2015) 5388-5396.

[3] M. Danelljan, G. Hager, F. Shahbaz Khan, M. Felsberg, Adaptive decontamination of the training set: A unified formulation for discriminative visual tracking, C. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. (2016) 1430-1438.

[4] Y. Li, J. Zhu, A scale adaptive kernel correlation filter tracker with feature integration, C. European Conference on Computer Vision. Springer International Publishing. (2014) 254-265.

[5] N.Y. Wang, D.Y. Yeung, Learning a deep compact image representation for visual tracking, C. Advances in neural information processing systems. (2013) 809-817.

[6] Y. LeCun, Y. Bengio, G. Hinton, Deep learning, J. Nature. 521.7553 (2015) 436-444.

[7] X. Feng, W. Mei, D. Hu, A Review of Visual Tracking with Deep Learning, J. (2016)

[8] N.Y. Wang, S.Y. Li, A. Gupta, Transferring Rich Feature Hierarchies for Robust Visual Tracking, J. arXiv preprint arXiv: 1501.04587 (2015)

[9] H. Nam, B. Han, Learning multi-domain convolutional neural networks for visual tracking, C. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. (2016) 4293-4302.

[10] H. Fan, H. Ling. SANet: Structure-Aware Network for Visual Tracking, J. arXiv preprint arXiv: 1611.06878 (2016).

[11] J. Choi, J. Kwon, K. M. Lee, Visual Tracking by Reinforced Decision Making, J. arXiv preprint arXiv:1702.06291 (2017).

(5)

[13] M. Kristan, J. Matas, A. Leonardis, et al, The visual object tracking vot2015 challenge results, C. Proceedings of the IEEE international conference on computer vision workshops. (2015) 1-23.

[14] M. Mueller, N. Smith, B. Ghanem, A benchmark and simulator for uav tracking, C. European Conference on Computer Vision. Springer International Publishing. (2016) 445-461.

[15] Y. Wu, J. Lim, and M. H. Yang, Object tracking benchmark, J. IEEE Transactions on Pattern Analysis and Machine Intelligence, 37.9 (2015) 1834-1848.

Figure

Figure 1. Precision and success plots on OTB100. The numbers in the legend indicate the representative precisions at 20 pixels for precision plots, and the area-under-curve scores for success plots

References

Related documents

Liver cell tumors have been reported in patients receiving long-term therapy with androgenic anabolic steroids in high doses (see WARNINGS).. Withdrawal of the drugs did not lead to

Multiphase model Multiphase model tumour cells host cells deformable and degradable ECM extracellular liquid DDDD DDDD... Multiphase models

In this pa- per, we propose a novel edit similarity learning approach ( GESL ) driven by the idea of ( ǫ, γ, τ )-goodness, which allows us to derive generalization guarantees using

Hence, we hypothesize that patients’ switching behavior across hospitals will be associated with an increase in duplicate tests due to lack of access to their

eRouter DOCSIS Embedded Router: An eSAFE that is compliant with [eRouter], providing IPv4 and/or IPv6 data forwarding, address configuration, and Domain Name services to

This chapter describes the process of making watertight pipe joints using heat and various filler metals.. After studying this chapter, you will be able to: • Identify the

Therefore, this study suggests that additional support services need to be provided to previously unsuccessful online learners, while students who succeed online should be

Commercial banks characterized with the lowest average TE score over the period under study included the Commercial Bank of Ethiopia, United Bank, and Buna International Bank