+ All Categories
Home > Documents > TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking...

TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking...

Date post: 06-Jun-2020
Category:
Upload: others
View: 7 times
Download: 0 times
Share this document with a friend
18
TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking in the Wild Matthias M¨ uller , Adel Bibi , Silvio Giancola , Salman Alsubaihi, and Bernard Ghanem King Abdullah University of Science and Technology, Thuwal, KSA {name.surname,matthias.mueller.2}@kaust.edu.sa http://www.tracking-net.org Abstract. Despite the numerous developments in object tracking, fur- ther improvement of current tracking algorithms is limited by small and mostly saturated datasets. As a matter of fact, data-hungry trackers based on deep-learning currently rely on object detection datasets due to the scarcity of dedicated large-scale tracking datasets. In this work, we present TrackingNet, the first large-scale dataset and benchmark for object tracking in the wild. We provide more than 30K videos with more than 14 million dense bounding box annotations. Our dataset covers a wide selection of object classes in broad and diverse context. By releas- ing such a large-scale dataset, we expect deep trackers to further improve and generalize. In addition, we introduce a new benchmark composed of 500 novel videos, modeled with a distribution similar to our training dataset. By sequestering the annotation of the test set and providing an online evaluation server, we provide a fair benchmark for future devel- opment of object trackers. Deep trackers fine-tuned on a fraction of our dataset improve their performance by up to 1.6% on OTB100 and up to 1.7% on TrackingNet Test. We provide an extensive benchmark on TrackingNet by evaluating more than 20 trackers. Our results suggest that object tracking in the wild is far from being solved. Keywords: Object Tracking, Dataset, Benchmark, Deep Learning 1 Introduction Object tracking is a common task in computer vision, with a long history span- ning decades [50,30,44]. Despite considerable progress in the field, object tracking remains a challenging task. Current trackers perform well on established datasets such as OTB [48,49] and VOT [25,26,27,24,22,23] benchmarks. However, most of these datasets are fairly small and do not fully represent the challenges faced when tracking objects in the wild. * This work was supported by the King Abdullah University of Science and Tech- nology (KAUST) Office of Sponsored Research (OSR). denote equal contribution.
Transcript
Page 1: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Dataset and

Benchmark for Object Tracking in the Wild ∗

Matthias Muller†, Adel Bibi†, Silvio Giancola†,Salman Alsubaihi, and Bernard Ghanem

King Abdullah University of Science and Technology, Thuwal, KSA{name.surname,matthias.mueller.2}@kaust.edu.sa

http://www.tracking-net.org

Abstract. Despite the numerous developments in object tracking, fur-ther improvement of current tracking algorithms is limited by small andmostly saturated datasets. As a matter of fact, data-hungry trackersbased on deep-learning currently rely on object detection datasets dueto the scarcity of dedicated large-scale tracking datasets. In this work,we present TrackingNet, the first large-scale dataset and benchmark forobject tracking in the wild. We provide more than 30K videos with morethan 14 million dense bounding box annotations. Our dataset covers awide selection of object classes in broad and diverse context. By releas-ing such a large-scale dataset, we expect deep trackers to further improveand generalize. In addition, we introduce a new benchmark composed of500 novel videos, modeled with a distribution similar to our trainingdataset. By sequestering the annotation of the test set and providing anonline evaluation server, we provide a fair benchmark for future devel-opment of object trackers. Deep trackers fine-tuned on a fraction of ourdataset improve their performance by up to 1.6% on OTB100 and upto 1.7% on TrackingNet Test. We provide an extensive benchmark onTrackingNet by evaluating more than 20 trackers. Our results suggestthat object tracking in the wild is far from being solved.

Keywords: Object Tracking, Dataset, Benchmark, Deep Learning

1 Introduction

Object tracking is a common task in computer vision, with a long history span-ning decades [50,30,44]. Despite considerable progress in the field, object trackingremains a challenging task. Current trackers perform well on established datasetssuch as OTB [48,49] and VOT [25,26,27,24,22,23] benchmarks. However, mostof these datasets are fairly small and do not fully represent the challenges facedwhen tracking objects in the wild.

∗This work was supported by the King Abdullah University of Science and Tech-nology (KAUST) Office of Sponsored Research (OSR).

† denote equal contribution.

Page 2: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

2 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

Fig. 1. Examples of tracking from our novel TrackingNet Test set.

Following the rise of deep learning in computer vision, the tracking commu-nity is currently embracing data-driven learning methods. Most trackers sub-mitted to the annual challenge VOT17 [23] use deep features, while they werenonexistent in earlier versions VOT13 [26] and VOT14 [27]. In addition, nine outof the ten top-performing trackers in VOT17 [23] rely on deep features, outper-forming the previous state-of-the-art trackers. However, the tracking communitystill lacks a dedicated large-scale dataset to train deep trackers. As a consequence,deep trackers are often restricted to using pretrained models from object clas-sification [6] or use object detection datasets such as ImageNet Videos [42]. Asan example of this, SiameseFC [2] and CFNet [45] show outstanding results bytraining specific Convolutional Neural Networks (CNNs) for tracking.

Since classical trackers rely on handcrafted features and because existingtracking datasets are small, there is currently no clear split between data usedfor training and testing. Recent benchmarks [23,35] now consider putting asidea sequestered test set to provide a fair comparison. Hence, it is common tosee trackers developed and trained on the OTB [49] dataset before compet-ing on VOT [25]. Note that VOT15 [24] is sampled from existing datasets likeOTB100 [49] and ALOV300 [43], resulting in overlapping sequences (e.g. basket-ball, car, singer, etc...). Even though the redundancy is contained, one needs tobe careful while selecting training video sequences, since training deep trackerson testing videos is not fair. As a result, there is usually not enough data to traindeep networks for tracking and data from different fields are used to pre-trainmodels, which is a limiting factor for certain architectures.

In this paper, we present TrackingNet, a large-scale object tracking datasetdesigned to train deep trackers. Our dataset has several advantages. First, thelarge training set enables the development of deep design specific for tracking.Second, the specificity of the dataset for object tracking enables novel architec-tures to focus on the temporal context between consecutive frames. Current large

Page 3: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Object Tracking Dataset 3

scale object detection datasets do not provide data densely annotated in time.Third, TrackingNet represents real-world scenarios by sampling over YouTubevideos. As such, TrackingNet videos contain a rich distribution of object classes,which we enforce to be shared between training and testing. Last, we evaluatetracker performance on a segregated testing set with a similar distribution overobject classes and motion. Trackers do not have access to the annotations ofthese videos but can obtain results and insights through an evaluation server.Contributions. (i) We present TrackingNet, the first large-scale dataset forobject tracking. We analyze the characteristics, attributes and uniqueness ofTrackingNet when compared with other datasets (Section 3). (ii) We provideinsights into different techniques to generate dense annotations from coarse ones.We show that most trackers can produce accurate and reliable dense annotationsover 1 second-long intervals. (Section 4). (iii) We provide an extended baselinefor state-of-the-art trackers benchmarked on TrackingNet. We show that pre-training deep models on TrackingNet can improve their performance on otherdatasets by increasing their metrics by up to 1.7%. (Section 5).

2 Related Work

In the following, we provide an overview of the various research on object track-ing. The tasks in the field can be clustered between multi-object tracking [49,25]and single-object tracking [28,35]. The former focuses on multiple instance track-ing of class-specific objects, relying on strong and fast object detection algo-rithms and association estimation between consecutive frames. The latter is thetarget of this work. It approaches the problem by tracking-by-detection, whichconsists of two main components: model representation, either generative [20,41]or discriminative [51,14], and object search, a trade-off between computationalcost and dense sampling of the region of interest.Correlation Filter Trackers. In recent years, correlation filter (CF) track-ers [4,19,16,1] have emerged as the most common, fastest and most accuratecategory of trackers. CF trackers learn a filter at the first frame, which repre-sents the object of interest. This filter localizes the target in successive framesbefore being updated. The main reason behind the impressive performance ofCF trackers lies in the approximate dense sampling achieved by circulantly shift-ing the target patch samples [19]. Also, the remarkable runtime performance isachieved by efficiently solving the underlying ridge regression problem in theFourier domain [4]. Since the inception of CF trackers with single-channel fea-tures [4,19], they have been extended with kernels [16], multi-channel features [9]and scale adaptation [32]. In addition, many works enhance the original formu-lation by adapting the regression target [3], adding context [12,37], spatiallyregularizing the learned filters and learning continuous filters [10].Deep Trackers. Beside the CF trackers that use deep features from objectdetection networks, few works explore more complete deep learning approaches.A first approach consists of learning generic features on a large-scale objectdetection dataset and successively fine-tuning domain-specific layers to be target-

Page 4: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

4 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

specific in an online fashion. MDNET [38] shows the success of such a methodby winning the VOT15 [24] challenge. A second approach consists of training afully convolutional network and using a feature map selection method to choosebetween shallow and deep layers during tracking [47]. The goal is to find a goodtrade-off between general semantic and more specific discriminative features, aswell as, to remove noisy and irrelevant feature maps.

While both of these approaches achieve state-of-the-art results, their compu-tation cost prohibits these algorithms from being deployed in real applications. Athird approach consists of using Siamese networks that predict motion betweenconsecutive frames. Such trackers are usually trained offline on a large-scaledataset using either deep regression [15] or a CNN matching function [2,45,13].Due to their simple architecture and lack of online fine-tuning, only a forwardpass has to be executed at test time. This results in very fast run-times (upto 100fps on a GPU) while achieving competitive accuracy. However, since themodel is not updated at test time, the accuracy highly depends on how well thetraining dataset captures appearance nuisances that occur while tracking variousobjects. Such approaches would benefit from a large-scale dataset like the onewe propose in this paper.

Object Tracking Datasets. Numerous datasets are available for object track-ing, the most common ones being OTB [49], VOT [25], ALOV300 [43] andTC128 [33] for single-object tracking and MOT [28,35] for multi-object track-ing. VIVID [5] is an early attempt to build a tracking dataset for surveillancepurposes. OTB50 [48] and OTB100 [49] provide 51 and 98 video sequences an-notated with 11 different attributes and upright bounding boxes for each frame.TC128 [33] comprises 129 videos, based on similar attributes and upright bound-ing boxes. ALOV300 [43] comprises 314 videos sequences labelled with 14 at-tributes. VOT [25] proposes several challenges with up to 60 video sequences. Itintroduced rotated bounding boxes as well as extensive studies on object track-ing annotations. VOT-TIR is a specific dataset from VOT focusing on ThermalInfraRed videos.NUS PRO [29] gathers an application-specific collection of 365videos for people and rigid object tracking. UAV123 and UAV20L [36] gatheranother application-specific collection of 123 videos and 20 long videos capturedfrom a UAV or generated from a flight simulator. NfS [11] provides a set of100 videos with high framerate, in an attempt to focus on fast motion. Table 1provides a detailed overview of the most popular tracking datasets.

Despite the availability of several datasets for object tracking, large scaledatasets are necessary to train deep trackers. Therefore, current deep track-ers rely on object detection datasets such as ImageNet Video [42] or Youtube-BoundingBoxes [40]. Those datasets provide object detection bounding boxeson videos, relatively sparse in time or at a low frame rate. Thus, they lack mo-tion information about the object dynamics in consecutive frames. Still, they arewidely used to pre-train deep trackers. They provide deep feature representationwith object knowledge that can be transferred from detection to tracking.

Page 5: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Object Tracking Dataset 5

Table 1. Comparison of current datasets for object tracking.

Datasets Nb Videos Nb Annot. Frame per Video Nb Classes

VIVID [5] 9 16274 1808.2 -TC128 [33] 129 55652 431.4 -OTB50 [48] 51 29491 578.3 -OTB100 [49] 98 58610 598.1 -VOT16 [22] 60 21455 357.6 -VOT17 [23] 60 21356 355.9 -UAV20L [36] 20 58670 2933.5 -UAV123 [36] 91 113476 1247.0 -NUS PRO [29] 365 135305 370.7 -ALOV300 [43] 314 151657 483.0 -NfS [13] 100 383000 3830.0 -MOT16 [35] 7 182326 845.6 -MOT17 [35] 21 564228 845.6 -

TrackingNet (Train) 30132 14205677 471.4 27TrackingNet (Test) 511 225589 441.5 27

3 TrackingNet

In this section, we introduce TrackingNet, a large-scale dataset for object track-ing. TrackingNet assembles a total of 30,643 video segments with an averageduration of 16.6s. All the 14,431,266 frames extracted from the 140 hours ofvisual content are annotated with a single upright bounding box. We provide acomparison with other tracking datasets in Table 1 and Figure 2.

Our work attempts to bridge the gap between data-hungry deep trackers andscarcely-available large scale datasets. Our proposed tracking dataset is largerthan the previous largest one by 2 orders of magnitude. We build TrackingNetto address object tracking in the wild. Therefore, the dataset copes with a largevariety of frame rates, resolutions, context and object classes. In contrast withprevious tracking datasets, TrackingNet is split between training and testing.We carefully select 30,132 training videos from Youtube-BoundingBoxes [40]and build a novel set of 511 testing videos with a distribution similar to thetraining set.

3.1 From YT-BB to TrackingNet Training Set

Youtube-BoundingBoxes (YT-BB) [40] is a large scale dataset for object detec-tion. This dataset consists of approximately 380,000 video segments, annotatedevery second with upright bounding boxes. Those videos are gathered directlyfrom YouTube, with a wide diversity in resolution, frame rate and duration.

Since YT-BB focuses on object detection, the object class is provided alongwith the bounding boxes. The dataset proposes a list of 23 object classes repre-sentative of the videos available on the YouTube platform. For the sake of track-ing, we remove the object classes that lack motion by definition, in particular

Page 6: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

6 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

Fig. 2. Comparison of tracking datasets distributed across the number of videos andthe average length of the videos. The size of circles is proportional to the number ofannotated bounding boxes. Our dataset has the largest amount of videos and framesand the video length is still reasonable for short video tracking.

potted plant and toilet. Since the person class represents 25% of the annotations,we split it into 7 different classes based on their context. Overall, the distributionof the object classes in TrackingNet is shown in Figure 3.

Fig. 3. Definition of object classes and macro classes.

To ensure decent quality in the videos for tracking purposes, we filteredout 90% of the the videos based on attribute criteria. First, we avoid smallsegments by removing videos shorter than 15 seconds. Second, we only consideredbounding boxes that covered less than 50% of the frame. Last, we preservesegments that contain at least a reasonable amount of motion between boundingboxes. During such filtering, we preserved the original distribution of the 21object classes provided by YT-BB, to prevent bias in the dataset. We end up

Page 7: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Object Tracking Dataset 7

with a training set of 30,132 videos, which we split into 12 training subsets, eachof which contains 2,511 videos and preserves the original YT-BB object classesdistribution.

Coarse annotations are provided by YT-BB at 1 fps. In order to increasethe annotation density, we rely on a mixture of state-of-the-art trackers to fillin missing annotations. We claim that any tracker is reliable on a small timelapse of 1 second. We present in Section 4 the performance of state-of-the-arttrackers on 1 second-long video segments from OTB100. As a result, we denselyannotated the 30,132 videos using a weighted average between a forward and abackward pass using the DCF tracker [16]. By doing so, we provide a denselyannotated training dataset for object tracking, along with code for automaticallydownloading videos from YouTube and extracting the annotated frames.

3.2 From YT-CC to TrackingNet Testing Set

Alongside the training dataset, we compile a novel dataset for testing, whichcomprises 511 videos from YouTube with Creative Commons licence, namely YT-CC. We carefully select those videos to reflect the object class distribution fromthe training set. We ensure that those videos do not contain any copyrights, sothey can be shared. We then used Amazon Mechanical Turk workers (Turkers) forannotating those videos. We annotate the first bounding boxes and define specificrules for the Turkers to carefully annotate the successive frames. We define theobjects as in YT-BB for object detection, i.e. with the smallest bounding boxfitting any visible part of the object to track.

Annotations should be defined in a deterministic way, using rules that areagreed upon and abided by during the annotation process. By defining the small-est upright bounding box around an object, we avoid any ambiguity. However,the bounding box may contain a large amount of background. For instance, thearm and the legs are always included for the person class, regardless of the per-son’s pose. We argue that a tracker should be able to cope with deformableobjects and to understand what it is tracking. In a similar fashion, the tails ofanimal are always included. In addition, the bounding box of an object is ad-justed as a function of its visibility in the frame. Estimating the position of anoccluded part of the object is not deterministic hence should be avoided. Forinstance, the handle of the object class knife could be hidden by the hand. Insuch cases, only the blade is annotated.

We use the VATIC tool [46] to annotate the frames. It incorporates an opticalflow algorithm to guess the position of the next bounding boxes in successiveframes. Turkers may annotate a non-tight bounding box around the object orrely on the optical flow to determine the bounding box location and size. Toavoid such behavior, we visually inspect every single frame after each annotationround, rewarding good Turkers and rejecting bad annotations. We either restartthe video annotation from scratch or ask Turkers to fine-tune previous results.With our supervision in the loop, we ensure the quality of our annotations aftera few iterations, discourage bad annotators and incentivize the good ones.

Page 8: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

8 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

3.3 Attributes

Successively, each video is annotated with a list of attributes defined in Table 2.15 attributes are provided for our testing set, the first 5 are extracted automat-ically by analyzing the variation of the bounding boxes in time while the last10 are manually checked by visually analyzing the 511 videos of our dataset.An overview of the attribute distribution is given in Figure 4 and compared toOTB100 [49] and VOT17 [23].

Table 2. List and description of the 15 attributes that characterize videos in Track-ingNet. Top: automatically estimated. Bottom: visually inspected.

Attr Description

SV Scale Variation: the ratio of bounding box area is outside the range [0.5, 2] after 1s.ARC Aspect Ratio Change: the ratio of bounding box aspect ratio is outside the range [0.5, 2] after 1s.

FM Fast Motion: the motion of the ground truth bounding box is larger than the size of the bounding box.LR Low Resolution: at least one ground truth bounding box has less than 1000 pixels.OV Out-of-View: some portion of the target leaves the camera field of view.

IV Illumination Variation: the illumination of the target changes significantly.CM Camera Motion: abrupt motion of the camera.MB Motion Blur: the target region is blurred due to the motion of target or camera.BC Background Clutter: the background near the target has similar appearance as the target.

SOB Similar Object: there are objects of similar shape or same type near the target.

DEF Deformation: non-rigid object deformation.IPR In-Plane Rotation: the target rotates in the image plane.OPR Out-of-Plane Rotation: the target rotates out of the image plane.POC Partial Occlusion: the target is partially occluded.FOC Full Occlusion: the target is fully occluded.

First, we claim to have better control over the number of frames per videoin our dataset, with a more contained variation with respect to other datasets.We argue that such contained length diversity is more suitable for training witha constant batch size. Second, the distribution of the bounding box resolution ismore diverse in TrackingNet, providing more diversity in the scale of the objectsto track. Third, we show that challenges in OTB100 [49] and VOT17 [23] focuson objects with slightly larger motion, while TrackingNet shows a more naturalmotion distribution over the fastest moving instances in YT-BB. Similar conclu-sions can be drawn from the distribution of the aspect ratio change attribute.Fourth, more than 30% of the OTB100 instances have a constant aspect ratio,while VOT17 shows a flatter distribution. Once again, we argue that Track-ingNet contains a more natural distribution of objects present in the wild. Last,we show statistics over the 15 attributes, which will be used to generate attributespecific tracking results in Section 5. Overall, we see that our sequestered testingset has an attribute distribution similar to that of our training set.

3.4 Evaluation

Annotation for the testing set should not be revealed to ensure a fair compar-ison between trackers. We thus evaluate the trackers through an online server.

Page 9: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Object Tracking Dataset 9

Fig. 4. (top to bottom, left to right): Distribution of the tracking videos in termof Video length, BB Resolution, Motion Change, Scale Variation and attributes distri-

bution for the main tracking datasets.

In a similar OTB100 fashion, we perform a One Pass Evaluation (OPE) andmeasure the success and precision of the trackers over the 511 videos. The suc-cess S is measured as the Intersection over Union (IoU) of the pixels betweenthe ground truth bounding boxes (BBgt) and the ones generated by the track-ers (BBtr). The trackers are ranked using the Area Under the Curve (AUC)measurement [49]. The precision P is usually measured as the distance in pixelsbetween the centers Cgt and Ctr of the ground truth and the tracker boundingbox, respectively. The trackers are ranked using this metric with a conventionalthreshold of 20 pixels.

Since the precision metric is sensitive to the resolution of the images and thesize of the bounding boxes, we propose a third metric Pnorm. We normalize theprecision over the size of the ground truth bounding box, following Eq. 1. Thetrackers are then ranked using the AUC for normalized precision between 0 and0.5. By substituting the original precision with the normalized one, we ensurethe consistency of the metrics across different scales of objects to track. However,for bounding boxes with similar scale, success and normalized precision are verysimilar and show how far an annotation is from another. Nevertheless, we arguethat they will differ in the case of different scales. For the sake of consistency,we provide results using precision, normalized precision and success.

S =|BBtr ∩BBgt|

|BBtr ∪BBgt|P = ‖Ctr − Cgt‖2

Pnorm = ‖W(

Ctr − Cgt)

‖2 W = diag(BBgtx , BBgt

y )

(1)

Page 10: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

10 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

4 Dataset Experiments

Since TrackingNet Training Set (∼ 30K videos) is compiled from the YT-BBdataset, it is originally annotated with bounding boxes every second. Whilesuch sparse annotations might be satisfactory for some vision tasks, e.g. ob-ject classification and detection, deep network based trackers rely on learningthe temporal evolution of bounding boxes over time. For instance, Siamese-likearchitectures [47,45] need to observe a large number of similar and dissimilarpatches of the same object. Unfortunately, manually extending YT-BB is notfeasible for such large number of frames. Thus, we have entertained the possi-bility of tracker-aided annotation to generate the missing dense bounding boxannotations arising between the sparsely occurring original YT-BB ones. State-of-the-art trackers not only achieve impressive performance on standard trackingbenchmarks, but they also perform well at high frame rates.

To assess such capability, we conducted four different experiments to decidewhich tracker would perform best in densely annotating OTB100 [49]. We choseamong the following trackers: ECO [6], CSRDCF [34], BACF [12], SiameseFC[2], STAPLECA [37], STAPLE [1], SRDCF [7], SAMF [31], CSK [17], KCF [18],DCF [18] and MOSSE [4]. To mimic the 1-second annotation in TrackingNetTraining Set, we assume that all videos of OTB100 are captured at 30fps andthe OTB100 dataset is split into 1916 smaller sequences of 30 frames. We evaluatethe previously highlighted trackers on the 1916 sequences of OTB100 by runningthem forward and backward through each sequence.

xtWG = wtx

tFW + (1− wt)x

tBK (2)

The results of both the forward and backward passes are then combined bydirectly averaging the two results and by generating the convex combination(weighted average) according to Eq. 2, where xt

FW, xt

BKand xt

WGare the track-

ing results at frame t for the forward pass, backward pass, and the weightedaverage respectively. We tested the linear, quadratic, cubic and exponential de-cay combinations for the weight wt. Note that the maximum sequence length is30, thus t ∈ [1, 30]. The weighted average gives more weight to the results of theforward pass for frames closer to the first frame and vice versa. Figure 5 alongwith Table 3 show that most trackers perform almost equally well with thebest performance upon using the weighted average strategy. Thereafter, sinceSTAPLECA [37] generates a reasonable accuracy with a frame rate of 30fps,we find it suitable for annotating the large training set in TrackingNet. We runSTAPLECA in both a forward and a backward pass where the results of both arelater combined in a weighted average using a linear decay fashion as describedin Eq. 2 using wt = (1− t/30).

5 Tracking Benchmark

In our benchmark, we compare a large variety of tracking algorithms that coverall common tracking principles. The majority of current state-of-the-art alogrithms

Page 11: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Object Tracking Dataset 11

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Suc

cess

rat

e

OPE Success plots on OTB100 - All Sequences

ECO [0.805] - 1.71fpsCSRDCF [0.801] - 8.82fpsSTAPLE

CA [0.799] - 30.6fps

BACF [0.795] - 20.4fpsSTAPLE [0.795] - 47.6fpsSRDCF [0.792] - 4.56fpsSAMF [0.784] - 15.6fpsCSK [0.776] - 169fpsSiameseFC [0.772] - 23.4fpsKCF [0.772] - 205fpsDCF [0.771] - 261fpsMOSSE [0.743] - 324fps

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Suc

cess

rat

e

OPE Success plots on OTB100 - All Sequences

ECO [0.809] - 1.55fpsCSRDCF [0.809] - 6.15fpsSTAPLE

CA [0.803] - 28.7fps

STAPLE [0.801] - 44.5fpsBACF [0.799] - 19.3fpsSRDCF [0.798] - 4.41fpsSAMF [0.789] - 14.7fpsKCF [0.780] - 204fpsDCF [0.779] - 338fpsCSK [0.778] - 209fpsSiameseFC [0.777] - 21.9fpsMOSSE [0.749] - 401fps

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Suc

cess

rat

e

OPE Success plots on OTB100 - All Sequences

ECO [0.843] - 1.55fpsCSRDCF [0.841] - 6.15fpsSTAPLE

CA [0.841] - 28.7fps

STAPLE [0.838] - 44.5fpsBACF [0.834] - 19.3fpsSRDCF [0.833] - 4.41fpsSAMF [0.829] - 14.7fpsKCF [0.826] - 204fpsDCF [0.825] - 338fpsCSK [0.825] - 209fpsSiameseFC [0.815] - 21.9fpsMOSSE [0.793] - 401fps

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

Suc

cess

rat

e

OPE Success plots on OTB100 - All Sequences

ECO [0.839] - 1.55fpsCSRDCF [0.838] - 6.15fpsSTAPLE

CA [0.837] - 28.7fps

STAPLE [0.835] - 44.5fpsBACF [0.829] - 19.3fpsSRDCF [0.828] - 4.41fpsSAMF [0.825] - 14.7fpsKCF [0.821] - 204fpsDCF [0.820] - 338fpsCSK [0.818] - 209fpsSiameseFC [0.810] - 21.9fpsMOSSE [0.781] - 401fps

Fig. 5. Tracking results of 12 trackers on the OT100 dataset after splitting it intosequences of length 30 frames. left to right: forward pass, backward pass, linear andexponential decay average as in Eq 2.

Table 3. Tracking results on the 1sec-long OTB100 dataset using different averaging.

OPE Success Forward Backward Average Linear Quadratic Cubic Exponential

Weight (wt) 1 0 0.5 (1− (t/30)i) e−0.05t

ECO 0.805 0.809 0.824 0.843 0.833 0.838 0.839DCF 0.771 0.779 0.799 0.825 0.813 0.820 0.820STAPLE CA 0.799 0.803 0.823 0.841 0.830 0.836 0.835

are based on discriminative correlation filters with handcrafted or deep features.We select trackers to cover a large set of combinations of features and ker-nels. MOSSE [4], CSK [19], DCF [16], KCF [16] use simple features and do notadapt to scale variations. DSST [9], SAMF [32], and STAPLE [1] use more so-phisticated features like Colornames and try to compensate for scale variations.We also include trackers that propose some kind of general framework to im-prove upon correlation filter tracking. These include SRDCF [8], SAMFAT [32],STAPLECA [37], BACF [12] and ECO-HC [6]. We include CFNet [45] and Siame-seFC [2] to represent CNN matching trackers and MEEM [51] and DLSSVM [39]for structured SVM-based trackers. Last, we include some baseline trackers suchas TLD [21], Struck [14], ASLA [20] and IVT [41] for reference. Table 4 summa-rizes the selected trackers along with their representation scheme, search method,runtime and a generic description.

5.1 State-of-the-art Benchmark on TrackingNet

Figure 6 shows the results on the complete dataset. Note that the highest scorefor any tracker is about 60% success rate compared to around 90% on OTB.The top performing tracker is MDNET [38] that trains in an online fashion andis, as a result, able to adapt best. However, this comes at the cost of a veryslow runtime. Next are CFNet [45] and SiameseFC [2] that benefit from beingtrained on a large-scale dataset (ImageNet Videos). However, as we show later,their performance can be further improved by using our training dataset.

Page 12: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

12 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

Table 4. Evaluated Trackers. Representation: PI - Pixel Intensity, HOG - Histogram ofOriented Gradients, CN - Color Names, CH - Color Histogram, GK - Gaussian Kernel,K - Keypoints, BP - Binary Pattern, SSVM - Structured Support Vector Machine.Search: PF - Particle Filter, RS - Random Sampling, DS - Dense Sampling.

Tracker Representation Search FPS Venue

ASLA[20] Sparse PF 2.13 CVPR’12IVT[41] PCA PF 11.7 IJCVIP’08Struck[14] SSVM, Haar RS 16.4 ICCV’11TLD[21] BP RS 22.9 PAMI’11

CSK[19] PI, GK DS 127 ECCV’12DCF[16] HOG DS 175 PAMI’15KCF[16] HOG, GK DS 119 PAMI’15MOSSE[4] PI DS 223 CVPR’10

DSST[9] PCA-HOG, PI DS 11.9 BMVC’14SAMF[32] PI, HOG, CN, GK DS 6.61 ECCVW’14STAPLE[1] HOG, CH DS 22.1 CVPR’16CSRDCF HOG, CN, PI DS 6.17 IJCV’18

SRDCF[8] HOG DS 3.17 ICCV’15BACF[12] HOG DS 12.1 ICCV’17ECO HC[6] HOG DS 21.2 CVPR’17SAMF AT[32] PI, HOG, CN, GK DS 2.1 ECCV’16STAPLE CA[37] HOG, CH DS 15.9 CVPR’17

CFNET[45] Deep DS 10.7 CVPR’17SiameseFC[2] Deep DS 11.6 ECCVW’16

MDNET[38] Deep RS 0.625 CVPR’16ECO[6] Deep DS 4.16 CVPR’17

MEEM[51] SSVM RS 7.57 ECCV’14DLSSVM[39] SSVM RS 5.59 CVPR’16

5.2 Real-Time Tracking

For many real applications, tracking is not very useful if it cannot be done atreal-time. Therefore, we conduct an experiment to evaluate how well trackerswould perform in more realistic settings where frames are skipped if a tracker istoo slow. We do this by subsampling the sequence based on each tracker’s speed.Figure 7 shows the results of this experiment across the complete dataset. Asexpected, most trackers that run below real-time degrade. In the worst case, thisdegradation can be as much as 50%, as is the case for Struck [14]. More recenttrackers, in particular deep learning ones, are much less affected. CFNet [45] forexample, does not degrade at all even though it only sees every third frame. Thisis probably due to the fact that it relies on a generic object matching functionthat was trained on a large-scale dataset.

Page 13: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Object Tracking Dataset 13

0 10 20 30 40 50Location error threshold

0

0.2

0.4

0.6

0.8

Pre

cisi

on

OPE Precision Plots on OTB100 - All SequencesECO [0.909] - 8.27fpsMDNET [0.885] - 0.903fpsECO

HC [0.841] - 29.6fps

STAPLECA

[0.810] - 35.1fps

MEEM [0.797] - 10.2fpsCSRDCF [0.794] - 9.02fpsSAMF

AT [0.789] - 6.11fps

SRDCF [0.788] - 4.51fpsSTAPLE [0.784] - 59.8fpsCFNET [0.769] - 13.1fpsDLSSVM [0.767] - 4.41fpsSiameseFC [0.765] - 21.7fpsSAMF [0.743] - 16.8fpsBACF [0.700] - 25.4fpsKCF [0.695] - 212fpsDSST [0.693] - 28.3fpsDCF [0.690] - 333fpsStruck [0.584] - 17.8fpsTLD [0.546] - 33.4fpsCSK [0.519] - 299fpsASLA [0.513] - 1.6fpsIVT [0.434] - 12.5fpsMOSSE [0.414] - 355fps

0 0.1 0.2 0.3 0.4 0.5Normalized distance error threshold

0

0.2

0.4

0.6

0.8

Nor

mal

ized

Pre

cisi

on

OPE Normalized Precision Plots on OTB100 - All SequencesECO [0.752] - 8.27fpsMDNET [0.742] - 0.903fpsECO

HC [0.687] - 29.6fps

STAPLECA

[0.679] - 35.1fps

CFNET [0.660] - 13.1fpsSRDCF [0.653] - 4.51fpsCSRDCF [0.653] - 9.02fpsSTAPLE [0.653] - 59.8fpsSAMF

AT [0.645] - 6.11fps

DLSSVM [0.623] - 4.41fpsSiameseFC [0.621] - 21.7fpsSAMF [0.617] - 16.8fpsMEEM [0.615] - 10.2fpsBACF [0.600] - 25.4fpsDSST [0.573] - 28.3fpsKCF [0.550] - 212fpsDCF [0.549] - 333fpsStruck [0.480] - 17.8fpsTLD [0.437] - 33.4fpsASLA [0.435] - 1.6fpsCSK [0.418] - 299fpsIVT [0.371] - 12.5fpsMOSSE [0.323] - 355fps

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.2

0.4

0.6

0.8

Suc

cess

rat

e

OPE Success Plots on OTB100 - All SequencesECO [0.687] - 8.27fpsMDNET [0.660] - 0.903fpsECO

HC [0.630] - 29.6fps

STAPLECA

[0.598] - 35.1fps

SRDCF [0.598] - 4.51fpsCFNET [0.588] - 13.1fpsCSRDCF [0.587] - 9.02fpsSTAPLE [0.579] - 59.8fpsSiameseFC [0.569] - 21.7fpsBACF [0.551] - 25.4fpsSAMF

AT [0.549] - 6.11fps

DLSSVM [0.540] - 4.41fpsMEEM [0.539] - 10.2fpsSAMF [0.535] - 16.8fpsKCF [0.477] - 212fpsDCF [0.475] - 333fpsDSST [0.470] - 28.3fpsStruck [0.429] - 17.8fpsASLA [0.415] - 1.6fpsTLD [0.406] - 33.4fpsCSK [0.382] - 299fpsIVT [0.319] - 12.5fpsMOSSE [0.311] - 355fps

0 10 20 30 40 50Location error threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Pre

cisi

on

OPE Precision Plots on TrackingNetTest - All Sequences

MDNET [0.565]CFNET [0.533]SiameseFC [0.533]ECO [0.492]CSRDCF [0.480]SAMF [0.477]ECO

HC [0.476]

STAPLE [0.470]STAPLE

CA [0.468]

BACF [0.461]DSST [0.460]SRDCF [0.455]SAMF

AT [0.447]

DCF [0.419]KCF [0.419]DLSSVM [0.418]ASLA [0.406]Struck [0.402]MEEM [0.386]CSK [0.368]IVT [0.336]MOSSE [0.326]TLD [0.292]

0 0.1 0.2 0.3 0.4 0.5Normalized distance error threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Nor

mal

ized

Pre

cisi

on

OPE Normalized Precision Plots on TrackingNetTest - All Sequences

MDNET [0.705]SiameseFC [0.663]CFNET [0.654]CSRDCF [0.622]ECO [0.618]ECO

HC [0.608]

STAPLECA

[0.605]

STAPLE [0.603]SAMF [0.598]DSST [0.588]BACF [0.580]SRDCF [0.573]DLSSVM [0.562]SAMF

AT [0.560]

DCF [0.548]KCF [0.546]MEEM [0.545]Struck [0.539]ASLA [0.536]CSK [0.503]IVT [0.460]MOSSE [0.442]TLD [0.438]

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Suc

cess

rat

e

OPE Success Plots on TrackingNetTest - All Sequences

MDNET [0.606]CFNET [0.578]SiameseFC [0.571]ECO [0.554]ECO

HC [0.541]

CSRDCF [0.534]STAPLE

CA [0.529]

STAPLE [0.528]BACF [0.523]SRDCF [0.521]SAMF [0.504]ASLA [0.478]SAMF

AT [0.472]

DLSSVM [0.470]DSST [0.464]MEEM [0.460]Struck [0.456]DCF [0.448]KCF [0.447]CSK [0.429]IVT [0.417]TLD [0.400]MOSSE [0.388]

Fig. 6. Benchmark results on OTB100 (top) and on TrackingNet (bottom).

0 10 20 30 40 50Location error threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Pre

cisi

on

OPE Precision Plots on TrackingNetTest@vfps - All Sequences

CFNET [0.548]SiameseFC [0.518]ECO

HC [0.474]

ECO [0.470]STAPLE

CA [0.464]

STAPLE [0.463]CSRDCF [0.453]BACF [0.444]DSST [0.432]SAMF [0.429]DLSSVM [0.429]MDNET [0.428]DCF [0.417]KCF [0.416]SRDCF [0.398]MEEM [0.395]SAMF

AT [0.370]

CSK [0.363]MOSSE [0.324]TLD [0.292]IVT [0.280]ASLA [0.217]Struck [0.203]

0 0.1 0.2 0.3 0.4 0.5Normalized distance error threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Nor

mal

ized

Pre

cisi

on

OPE Normalized Precision Plots on TrackingNetTest@vfps - All Sequences

CFNET [0.665]SiameseFC [0.652]ECO

HC [0.606]

ECO [0.604]STAPLE

CA [0.603]

STAPLE [0.600]CSRDCF [0.587]DLSSVM [0.578]MDNET [0.574]BACF [0.566]DSST [0.560]MEEM [0.556]SAMF [0.548]DCF [0.545]KCF [0.544]SRDCF [0.524]CSK [0.500]SAMF

AT [0.498]

MOSSE [0.440]TLD [0.438]IVT [0.407]ASLA [0.355]Struck [0.350]

0 0.2 0.4 0.6 0.8 1Overlap Threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Suc

cess

Rat

e

OPE Success Plots on TrackingNetTest@vfps - All Sequences

CFNET [0.580]SiameseFC [0.559]ECO

HC [0.542]

ECO [0.539]STAPLE [0.530]STAPLE

CA [0.529]

BACF [0.512]CSRDCF [0.508]MDNET [0.498]DLSSVM [0.479]SAMF [0.476]SRDCF [0.474]MEEM [0.468]DSST [0.450]DCF [0.450]KCF [0.449]SAMF

AT [0.443]

CSK [0.431]TLD [0.400]MOSSE [0.391]IVT [0.379]Struck [0.345]ASLA [0.344]

Fig. 7. Benchmark results on TrackingNet with variable frame rate (tracker fps).

5.3 Retraining on TrainingNet

We fine-tune SiameseFC [2] on a fraction of TrackingNet to show how our datacan improve the tracking performance of deep-learning based trackers. The re-sults are shown in Table 5. By training on only one of the twelve chunks (2511videos) of our training dataset, we observe an increase in all the metrics onTrackingNet Test and OTB100. Fine-tuning using more chunks is expected toimprove the performance even further.

Table 5. Fine-tuning results for SiameseFC on OTB100 and TrackingNet Test.

Benchmark OTB100 TrackingNet Test

Metric Precision Norm. Prec. Success Precision NormPrec Success

SiameseFC (original) 0.765 0.621 0.569 0.533 0.663 0.571SiameseFC (fine-tuned) 0.781 0.632 0.576 0.543 0.673 0.581

Page 14: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

14 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

5.4 Attribute Specific Results

Each video in TrackingNet Test is annotated with 15 attributes described inSection 3. We evaluate all trackers per attribute to get insights about challengesfacing state-of-the-art tracking algorithms. We show the most interesting resultsin Figure 8 and refer the reader to the supplementary material for the re-maining attributes. We find that videos with in-plane rotation, low resolutiontargets, and full occlusion are consistently the most difficult. Trackers are leastaffected by illumination variation, partial occlusion, and object deformation.

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Suc

cess

rat

e

OPE Success plots on TrackingNetTest - In-Plane Rotation (56)

MDNET [0.469]SiameseFC [0.423]CFNET [0.414]ECO [0.406]CSRDCF [0.379]ECO

HC [0.379]

SAMF [0.365]SRDCF [0.364]DLSSVM [0.357]STAPLE

CA [0.354]

MEEM [0.348]Struck [0.346]STAPLE [0.344]BACF [0.340]SAMF

AT [0.339]

ASLA [0.328]DSST [0.323]CSK [0.317]KCF [0.317]DCF [0.316]TLD [0.310]IVT [0.276]MOSSE [0.276]

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Suc

cess

rat

e

OPE Success plots on TrackingNetTest - Low Resolution (49)

CFNET [0.514]ECO [0.468]MDNET [0.462]SiameseFC [0.455]ECO

HC [0.432]

SRDCF [0.429]STAPLE

CA [0.428]

STAPLE [0.419]BACF [0.401]CSRDCF [0.387]ASLA [0.378]SAMF

AT [0.361]

SAMF [0.336]DLSSVM [0.334]DSST [0.320]MEEM [0.320]TLD [0.304]Struck [0.302]CSK [0.281]DCF [0.268]KCF [0.268]MOSSE [0.253]IVT [0.235]

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Suc

cess

rat

e

OPE Success plots on TrackingNetTest - Full Occlusion (24)

MDNET [0.494]ECO [0.467]SRDCF [0.454]ECO

HC [0.449]

CFNET [0.446]SAMF [0.438]CSRDCF [0.432]STAPLE [0.431]STAPLE

CA [0.426]

SAMFAT

[0.410]

DLSSVM [0.403]SiameseFC [0.403]BACF [0.398]KCF [0.394]DCF [0.392]ASLA [0.385]MEEM [0.384]DSST [0.381]Struck [0.356]CSK [0.349]TLD [0.317]IVT [0.314]MOSSE [0.276]

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Suc

cess

rat

e

OPE Success plots on TrackingNetTest - Partial Occlusion (238)

MDNET [0.594]CFNET [0.555]SiameseFC [0.546]ECO [0.523]ECO

HC [0.515]

CSRDCF [0.505]STAPLE

CA [0.504]

STAPLE [0.499]SRDCF [0.499]BACF [0.496]SAMF [0.495]DLSSVM [0.459]SAMF

AT [0.458]

MEEM [0.456]ASLA [0.455]DSST [0.451]Struck [0.443]DCF [0.443]KCF [0.443]CSK [0.415]IVT [0.403]TLD [0.385]MOSSE [0.368]

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Suc

cess

rat

e

OPE Success plots on TrackingNetTest - Deformation (291)

MDNET [0.592]SiameseFC [0.557]CFNET [0.551]ECO [0.520]CSRDCF [0.520]ECO

HC [0.508]

STAPLECA

[0.504]

STAPLE [0.503]BACF [0.496]SAMF [0.481]SRDCF [0.470]SAMF

AT [0.470]

DLSSVM [0.462]ASLA [0.460]DSST [0.459]MEEM [0.453]Struck [0.446]KCF [0.434]DCF [0.434]CSK [0.426]IVT [0.391]MOSSE [0.377]TLD [0.351]

0 0.2 0.4 0.6 0.8 1Overlap threshold

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Suc

cess

rat

e

OPE Success plots on TrackingNetTest - Illumination Variation (53)

MDNET [0.626]ECO [0.588]CFNET [0.580]SiameseFC [0.574]ECO

HC [0.570]

SRDCF [0.553]STAPLE [0.547]STAPLE

CA [0.547]

CSRDCF [0.546]SAMF [0.541]ASLA [0.505]BACF [0.497]SAMF

AT [0.496]

DSST [0.484]DLSSVM [0.482]DCF [0.472]KCF [0.471]MEEM [0.469]Struck [0.459]CSK [0.427]IVT [0.396]MOSSE [0.389]TLD [0.387]

Fig. 8. Per-attribute results on TrackingNet Test.

6 Conclusion

In this work, we present TrackingNet, which is, to the best of our knowledge, thelargest dataset for object tracking. We show how large-scale existing datasets forobject detection can be leveraged for object tracking by a novel interpolationmethod. We also benchmark more than 20 tracking algorithms on this noveldataset and shed light on what attributes are especially difficult for currenttrackers. Lastly, we verify the usefulness of our large dataset in improving theperformance of some deep learning based trackers.

In the future, we aim to extend the test set from 500 to 1000 videos. Weplan to sample the extra 500 videos from different classes within the same cat-egory (e.g. tortoise / animal). This will allow for further evaluation in regardsto generalization. After publication, we plan to release the training set withour interpolated annotations. We will also release the test sequences with ini-tial bounding box annotations and the corresponding integration for the OTBtoolkit. At the same time, we will publish our online evaluation server to allowresearches to rank their tracking algorithms instantly.

Page 15: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Object Tracking Dataset 15

References

1. Bertinetto, L., Valmadre, J., Golodetz, S., Miksik, O., Torr, P.H.: Staple: Comple-mentary learners for real-time tracking. In: Proceedings of the IEEE Conferenceon Computer Vision and Pattern Recognition. pp. 1401–1409 (2016)

2. Bertinetto, L., Valmadre, J., Henriques, J.F., Vedaldi, A., Torr, P.H.: Fully-convolutional siamese networks for object tracking. In: European conference oncomputer vision. pp. 850–865. Springer (2016)

3. Bibi, A., Mueller, M., Ghanem, B.: Target response adaptation for correlationfilter tracking. In: European conference on computer vision. pp. 419–433. Springer(2016)

4. Bolme, D.S., Beveridge, J.R., Draper, B.A., Lui, Y.M.: Visual object track-ing using adaptive correlation filters. In: Computer Vision and Pattern Recog-nition (CVPR), 2010 IEEE Conference on. pp. 2544–2550 (June 2010).https://doi.org/10.1109/CVPR.2010.5539960

5. Collins, R., Zhou, X., Teh, S.K.: An open source tracking testbed and evaluationweb site. In: IEEE International Workshop on Performance Evaluation of Trackingand Surveillance (PETS 2005), January 2005 (January 2005)

6. Danelljan, M., Bhat, G., Khan, F.S., Felsberg, M.: Eco: efficient convolution op-erators for tracking. In: Proceedings of the 2017 IEEE Conference on ComputerVision and Pattern Recognition (CVPR), Honolulu, HI, USA. pp. 21–26 (2017)

7. Danelljan, M., Hager, G., Shahbaz Khan, F., Felsberg, M.: Learning spatially reg-ularized correlation filters for visual tracking. In: Proceedings of the IEEE Inter-national Conference on Computer Vision. pp. 4310–4318 (2015)

8. Danelljan, M., Hager, G., Shahbaz Khan, F., Felsberg, M.: Learning spatially regu-larized correlation filters for visual tracking. In: The IEEE International Conferenceon Computer Vision (ICCV) (Dec 2015)

9. Danelljan, M., Hger, G., Shahbaz Khan, F., Felsberg, M.: Accurate scale estimationfor robust visual tracking. In: Proceedings of the British Machine Vision Confer-ence. BMVA Press (2014). https://doi.org/http://dx.doi.org/10.5244/C.28.65

10. Danelljan, M., Robinson, A., Shahbaz Khan, F., Felsberg, M.: Beyond correlationfilters: Learning continuous convolution operators for visual tracking. In: ECCV(2016)

11. Galoogahi, H.K., Fagg, A., Huang, C., Ramanan, D., Lucey, S.: Need for speed: Abenchmark for higher frame rate object tracking. arXiv preprint arXiv:1703.05884(2017)

12. Galoogahi, H.K., Fagg, A., Lucey, S.: Learning background-aware correlation filtersfor visual tracking. In: Proceedings of the 2017 IEEE Conference on ComputerVision and Pattern Recognition (CVPR), Honolulu, HI, USA. pp. 21–26 (2017)

13. Guo, Q., Feng, W., Zhou, C., Huang, R., Wan, L., Wang, S.: Learning dynamicsiamese network for visual object tracking. In: The IEEE International Conferenceon Computer Vision (ICCV) (Oct 2017)

14. Hare, S., Saffari, A., Torr, P.H.S.: Struck: Structured output tracking with kernels.In: 2011 International Conference on Computer Vision. pp. 263–270. IEEE (Nov2011). https://doi.org/10.1109/ICCV.2011.6126251

15. Held, D., Thrun, S., Savarese, S.: Learning to track at 100 fps with deep regressionnetworks. In: European Conference Computer Vision (ECCV) (2016)

16. Henriques, J.F., Caseiro, R., Martins, P., Batista, J.: High-speed tracking with ker-nelized correlation filters. Pattern Analysis and Machine Intelligence, IEEE Trans-actions on (2015). https://doi.org/10.1109/TPAMI.2014.2345390

Page 16: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

16 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

17. Henriques, J.F., Caseiro, R., Martins, P., Batista, J.: Exploiting the circulant struc-ture of tracking-by-detection with kernels. In: European conference on computervision. pp. 702–715. Springer (2012)

18. Henriques, J.F., Caseiro, R., Martins, P., Batista, J.: High-speed tracking withkernelized correlation filters. IEEE Transactions on Pattern Analysis and MachineIntelligence 37(3), 583–596 (2015)

19. Henriques, J., Caseiro, R., Martins, P., Batista, J.: Exploiting the circulant struc-ture of tracking-by-detection with kernels. In: Fitzgibbon, A., Lazebnik, S., Perona,P., Sato, Y., Schmid, C. (eds.) Computer Vision ECCV 2012, Lecture Notes inComputer Science, vol. 7575, pp. 702–715. Springer Berlin Heidelberg (2012)

20. Jia, X., Lu, H., Yang, M.H.: Visual tracking via adaptive structural lo-cal sparse appearance model. In: Computer Vision and Pattern Recog-nition (CVPR), 2012 IEEE Conference on. pp. 1822–1829 (June 2012).https://doi.org/10.1109/CVPR.2012.6247880

21. Kalal, Z., Mikolajczyk, K., Matas, J.: Tracking-Learning-Detection. IEEE trans-actions on pattern analysis and machine intelligence 34(7), 1409–1422 (Dec 2011).https://doi.org/10.1109/TPAMI.2011.239

22. Kristan, M., Leonardis, A., Matas, J., Felsberg, M., Pflugfelder, R., Cehovin,L., Vojir, T., Hager, G., Lukezic, A., Fernandez, G.: The visual object track-ing vot2016 challenge results. Springer (Oct 2016), http://www.springer.com/gp/book/9783319488806

23. Kristan, M., Leonardis, A., Matas, J., Felsberg, M., Pflugfelder, R.,Cehovin Zajc, L., Vojir, T., Hager, G., Lukezic, A., Eldesokey, A., Fer-nandez, G.: The visual object tracking vot2017 challenge results (2017),http://openaccess.thecvf.com/content ICCV 2017 workshops/papers/w28/Kristan The Visual Object ICCV 2017 paper.pdf

24. Kristan, M., Matas, J., Leonardis, A., Felsberg, M., Cehovin, L., Fernandez, G., Vo-jir, T., Hager, G., Nebehay, G., Pflugfelder, R.: The visual object tracking vot2015challenge results. In: Visual Object Tracking Workshop 2015 at ICCV2015 (Dec2015)

25. Kristan, M., Matas, J., Leonardis, A., Vojir, T., Pflugfelder, R., Fernan-dez, G., Nebehay, G., Porikli, F., Cehovin, L.: A novel performance eval-uation methodology for single-target trackers. IEEE Transactions on Pat-tern Analysis and Machine Intelligence 38(11), 2137–2155 (Nov 2016).https://doi.org/10.1109/TPAMI.2016.2516982

26. Kristan, M., Pflugfelder, R., Leonardis, A., Matas, J., Porikli, F., Cehovin, L.,Nebehay, G., Fernandez, G., Vojir, T., Gatt, A., et al.: The visual object trackingvot2013 challenge results. In: Computer Vision Workshops (ICCVW), 2013 IEEEInternational Conference on. pp. 98–111. IEEE (2013)

27. Kristan, M., Pflugfelder, R., Leonardis, A., Matas, J., Cehovin, L., Nebehay, G.,Vojir, T., Fernandez, G., Lukezic, A.: The visual object tracking vot2014 challengeresults (2014), http://www.votchallenge.net/vot2014/program.html

28. Leal-Taixe, L., Milan, A., Reid, I., Roth, S., Schindler, K.: Motchallenge 2015:Towards a benchmark for multi-target tracking. arXiv preprint arXiv:1504.01942(2015)

29. Li, A., Lin, M., Wu, Y., Yang, M.H., Yan, S.: Nus-pro: A new visual trackingchallenge. IEEE Transactions on Pattern Analysis and Machine Intelligence 38(2),335–349 (Feb 2016). https://doi.org/10.1109/TPAMI.2015.2417577

30. Li, X., Hu, W., Shen, C., Zhang, Z., Dick, A., Hengel, A.V.D.: A survey of appear-ance models in visual object tracking. ACM transactions on Intelligent Systemsand Technology (TIST) 4(4), 58 (2013)

Page 17: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

TrackingNet: A Large-Scale Object Tracking Dataset 17

31. Li, Y., Zhu, J.: A scale adaptive kernel correlation filter tracker with feature in-tegration. In: European Conference on Computer Vision. pp. 254–265. Springer(2014)

32. Li, Y., Zhu, J.: A scale adaptive kernel correlation filter tracker with feature in-tegration. In: Agapito, L., Bronstein, M.M., Rother, C. (eds.) Computer Vision- ECCV 2014 Workshops. pp. 254–265. Springer International Publishing, Cham(2015)

33. Liang, P., Blasch, E., Ling, H.: Encoding color information for visual tracking:Algorithms and benchmark. Image Processing, IEEE . . . pp. 1–14 (2015), http://ieeexplore.ieee.org/xpls/abs all.jsp?arnumber=7277070

34. Lukezic, A., Vojır, T., Zajc, L.C., Matas, J., Kristan, M.: Discriminative correlationfilter with channel and spatial reliability. In: Proceedings of the IEEE Conferenceon Computer Vision and Pattern Recognition. vol. 2 (2017)

35. Milan, A., Leal-Taixe, L., Reid, I., Roth, S., Schindler, K.: MOT16: A benchmarkfor multi-object tracking. arXiv:1603.00831 [cs] (Mar 2016), http://arxiv.org/abs/1603.00831, arXiv: 1603.00831

36. Mueller, M., Smith, N., Ghanem, B.: A benchmark and simulator for uav tracking.In: Proc. of the European Conference on Computer Vision (ECCV) (2016)

37. Mueller, M., Smith, N., Ghanem, B.: Context-aware correlation filter tracking. In:Proc. IEEE Conf. Comput. Vis. Pattern Recognit.(CVPR). pp. 1396–1404 (2017)

38. Nam, H., Han, B.: Learning multi-domain convolutional neural networks for visualtracking. In: The IEEE Conference on Computer Vision and Pattern Recognition(CVPR) (June 2016)

39. Ning, J., Yang, J., Jiang, S., Zhang, L., Yang, M.H.: Object tracking via dual linearstructured svm and explicit feature map. In: Proceedings of the IEEE conferenceon computer vision and pattern recognition. pp. 4266–4274 (2016)

40. Real, E., Shlens, J., Mazzocchi, S., Pan, X., Vanhoucke, V.: Youtube-boundingboxes: A large high-precision human-annotated data set for object detec-tion in video. In: 2017 IEEE Conference on Computer Vision and Pattern Recog-nition (CVPR). pp. 7464–7473. IEEE (2017)

41. Ross, D., Lim, J., Lin, R.S., Yang, M.H.: Incremental learning for robust vi-sual tracking. International Journal of Computer Vision 77(1-3), 125–141 (2008).https://doi.org/10.1007/s11263-007-0075-7

42. Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z.,Karpathy, A., Khosla, A., Bernstein, M., et al.: Imagenet large scale visual recog-nition challenge. International Journal of Computer Vision 115(3), 211–252 (2015)

43. Smeulders, A.W.M., Chu, D.M., Cucchiara, R., Calderara, S., Dehghan, A.,Shah, M.: Visual tracking: An experimental survey. IEEE Transactions onPattern Analysis and Machine Intelligence 36(7), 1442–1468 (July 2014).https://doi.org/10.1109/TPAMI.2013.230

44. Smeulders, A.W., Chu, D.M., Cucchiara, R., Calderara, S., Dehghan, A., Shah, M.:Visual tracking: An experimental survey. IEEE transactions on pattern analysisand machine intelligence 36(7), 1442–1468 (2014)

45. Valmadre, J., Bertinetto, L., Henriques, J., Vedaldi, A., Torr, P.H.: End-to-endrepresentation learning for correlation filter based tracking. In: Computer Visionand Pattern Recognition (CVPR), 2017 IEEE Conference on. pp. 5000–5008. IEEE(2017)

46. Vondrick, C., Patterson, D., Ramanan, D.: Efficiently scaling up crowdsourcedvideo annotation. International Journal of Computer Vision 101(1), 184–204(2013)

Page 18: TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking …openaccess.thecvf.com/content_ECCV_2018/papers/Matthias... · 2018-08-28 · TrackingNet: A Large-Scale Dataset

18 M. Muller, A. Bibi, S. Giancola, S. Alsubaihi and B. Ghanem

47. Wang, L., Ouyang, W., Wang, X., Lu, H.: Visual tracking with fully convolutionalnetworks. In: 2015 IEEE International Conference on Computer Vision (ICCV).pp. 3119–3127 (Dec 2015). https://doi.org/10.1109/ICCV.2015.357

48. Wu, Y., Lim, J., Yang, M.H.: Online object tracking: A benchmark. In: Computervision and pattern recognition (CVPR), 2013 IEEE Conference on. pp. 2411–2418.Ieee (2013)

49. Wu, Y., Lim, J., Yang, M.H.: Object tracking benchmark. IEEE Transactions onPattern Analysis and Machine Intelligence 37(9), 1834–1848 (2015)

50. Yilmaz, A., Javed, O., Shah, M.: Object tracking: A survey. Acm computing sur-veys (CSUR) 38(4), 13 (2006)

51. Zhang, J., Ma, S., Sclaroff, S.: MEEM: robust tracking via multiple experts usingentropy minimization. In: Proc. of the European Conference on Computer Vision(ECCV) (2014)


Recommended