Online object tracking A benchmark
- 格式:pdf
- 大小:5.70 MB
- 文档页数:8
2017目标跟踪算法综述2017目标跟踪算法综述时间:2017年7月31日。
本文所提的跟踪主要指的是单目标跟踪,多目标跟踪暂时不作为考虑范围。
本文主要从常用的评价标准,如EAO,EFO,fps等,分析2016-2017年最新出的目标跟踪文章在应用层面的优缺点。
算法原理不作为本文重点,旨在对比不同方法的“效果和时间性能”,方便大家根据不同的业务常见选择不同的方法。
本文按照以下几类进行展开,并不局限于传统方法或者深度学习。
TCNN C-COT ECO (根据名次递增)CFNet DCFNet SANet DRLT (端到端CNN RNN)CA-CF BACF (利用背景信息,框架性通用改进)ACET Deep-LK (优于C-COT or MDNet, 且CPU实时)LMCF Vision-based (速度提升,但性能提升不明显CPU实时)2017目标跟踪算法综述目标跟踪VOT2016 BenchMark评价标准介绍推荐网站TCNN C-COT ECO 根据名次递增TCNN17_arXiv CVPR_TCNN_Modeling and Propagating CNNs in a Tree Structure for Visual TrackingC-COT16_ECCV_CCOT_Beyond Correlation Filters - Learning ContinuousCFNet DCFNet SANet DRLT 端到端CNN RNNDCFNet17_arXiv prePrint_DCFNet_ Discriminant Correlation Filters Network for Visual TrackingSANet17_CVPR_SANet Structure-Aware Network for Visual Tracking DRLT17_arXiv prePrint_DRLT_Deep Reinforcement Learning for Visual Object Tracking in VideosCA-CF BACF 利用背景信息框架性通用改进CA-CF17_CVPR_CA-CFContext-Aware Correlation Filter Tracking BACF17_CVPR_BACF_Learning Background-Aware Correlation Filters for Visual TrackingACET Deep-LK 优于C-COT or MDNet 且CPU实时ACET17_arXiv prePrint_ACET_Active Collaborative Ensemble TrackingDeep-LK17_arXiv prePrint_Deep-LK_ for Efficient Adaptive Object TrackingLMCF Vision-based 速度提升但性能提升不明显CPU实时LMCF17_CVPR_LMCF_Large Margin Object Tracking with Circulant Feature MapsVision-based17_arXiv prePrint_NULL_Vision-based Real-Time Aerial Object Localization and Tracking for UAV Sensing System目标跟踪VOT2016 BenchMark评价标准介绍见目标跟踪VOT2016 BenchMark评价标准介绍推荐网站OTB Results: 这是foolwood总结的目标跟踪发展主线图:这是foolwood总结的这是浙江大学Mengmeng Wang同学在极视角公开课上展示的一个总结: foolwood这个人在github上最新的一些跟踪方法常见数据集上的结果比较。
benchmark方法-回复什么是benchmark方法?Benchmark方法是一种常用的研究方法,用于比较和评估不同实体、系统或策略之间的性能差异。
它是一种定量研究方法,通过设定标准或基准来衡量被测试对象的表现。
Benchmark方法广泛应用于各个领域,包括金融、科技、医疗等,用于评估各种产品、系统或策略的效果和可行性。
Benchmark方法的重要性和意义是什么?Benchmark方法的目的是提供一个客观的标准来评估不同实体、系统或策略的性能差异。
通过使用统一的基准,我们可以更好地识别出行业或领域内的佼佼者,并找到并改进低效或不理想的方案。
Benchmark方法为决策者提供了有力的工具和数据,帮助他们做出明智的决策和策略选择。
Benchmark方法的步骤是什么?1. 确定研究目的和对象:首先,我们需要明确研究的目的是什么,以及要测试或比较的实体、系统或策略是什么。
例如,我们可能想要比较两种不同的投资组合的回报率,或者评估两个电子产品的性能差异。
2. 设定基准或标准:接下来,我们需要确定一个标准或基准来比较被测试对象的性能。
这个基准应该是一个可衡量的、普遍接受的标准,以便于与其他研究结果进行比较和评估。
3. 收集数据和测量指标:在进行比较之前,我们需要收集和记录需要比较的数据。
这些数据应该是客观的、可衡量的,并与所设定的基准相对应。
比如,在比较两种投资组合的回报率时,我们需要收集和记录每个投资组合的历史数据。
4. 进行比较和分析:一旦数据收集完毕,我们可以开始进行比较和分析,将被测试对象的性能与所设定的基准进行对比。
这可以通过统计方法、图表或其他分析工具来完成。
比如,我们可以计算投资组合的年化收益率,并将其与市场指数进行比较。
5. 做出决策和改进:通过比较和分析的结果,我们可以得出关于被测试对象的性能和差异的结论。
根据这些结论,我们可以做出相应的决策和改进。
例如,在投资领域,如果一个投资组合的回报率明显低于基准,我们可能会考虑调整投资策略或选择其他的投资组合。
java英⽂参考⽂献java英⽂参考⽂献汇编 导语:Java是⼀门⾯向对象编程语⾔,不仅吸收了C++语⾔的各种优点,还摒弃了C++⾥难以理解的多继承、指针等概念,因此Java语⾔具有功能强⼤和简单易⽤两个特征。
下⾯⼩编为⼤家带来java英⽂参考⽂献,供各位阅读和参考。
java英⽂参考⽂献⼀: [1]Irene Córdoba-Sánchez,Juan de Lara. Ann: A domain-specific language for the effective design and validation of Java annotations[J]. Computer Languages, Systems & Structures,2016,:. [2]Marcelo M. Eler,Andre T. Endo,Vinicius H.S. Durelli. An Empirical Study to Quantify the Characteristics of Java Programs that May Influence Symbolic Execution from a Unit Testing Perspective[J]. The Journal of Systems & Software,2016,:. [3]Kebo Zhang,Hailing Xiong. A new version of code Java for 3D simulation of the CCA model[J]. Computer Physics Communications,2016,:. [4]S. Vidal,A. Bergel,J.A. Díaz-Pace,C. Marcos. Over-exposed classes in Java: An empirical study[J]. Computer Languages, Systems & Structures,2016,:. [5]Zeinab Iranmanesh,Mehran S. Fallah. Specification and Static Enforcement of Scheduler-Independent Noninterference in a Middleweight Java[J]. Computer Languages, Systems & Structures,2016,:. [6]George Gabriel Mendes Dourado,Paulo S Lopes De Souza,Rafael R. Prado,Raphael Negrisoli Batista,Simone R.S. Souza,Julio C. Estrella,Sarita M. Bruschi,Joao Lourenco. A Suite of Java Message-Passing Benchmarks to Support the Validation of Testing Models, Criteria and Tools[J]. Procedia Computer Science,2016,80:. [7]Kebo Zhang,Junsen Zuo,Yifeng Dou,Chao Li,Hailing Xiong. Version 3.0 of code Java for 3D simulation of the CCA model[J]. Computer Physics Communications,2016,:. [8]Simone Hanazumi,Ana C.~V. de Melo. A Formal Approach to Implement Java Exceptions in Cooperative Systems[J]. The Journal of Systems & Software,2016,:. [9]Lorenzo Bettini,Ferruccio Damiani. Xtraitj : Traits for the Java Platform[J]. The Journal of Systems & Software,2016,:. [10]Oscar Vega-Gisbert,Jose E. Roman,Jeffrey M. Squyres. Design and implementation of Java bindings in OpenMPI[J]. Parallel Computing,2016,:. [11]Stefan Bosse. Structural Monitoring with Distributed-Regional and Event-based NN-Decision Tree Learning Using Mobile Multi-Agent Systems and Common Java Script Platforms[J]. Procedia Technology,2016,26:. [12]Pablo Piedrahita-Quintero,Carlos Trujillo,Jorge Garcia-Sucerquia. JDiffraction : A GPGPU-accelerated JAVA library for numerical propagation of scalar wave fields[J]. Computer Physics Communications,2016,:. [13]Abdelhak Mesbah,Jean-Louis Lanet,Mohamed Mezghiche. Reverse engineering a Java Card memory management algorithm[J]. Computers & Security,2017,66:. [14]G. Bacci,M. Bazzicalupo,A. Benedetti,A. Mengoni. StreamingTrim 1.0: a Java software for dynamic trimming of 16S rRNA sequence data from metagenetic studies[J]. Mol Ecol Resour,2014,14(2):. [15]Qing‐Wei Xu,Johannes Griss,Rui Wang,Andrew R. Jones,Henning Hermjakob,Juan Antonio Vizcaíno. jmzTab: A Java interface to the mzTab data standard[J]. Proteomics,2014,14(11):. [16]Rody W. J. Kersten,Bernard E. Gastel,Olha Shkaravska,Manuel Montenegro,Marko C. J. D. Eekelen. ResAna: a resource analysis toolset for (real‐time) JAVA[J]. Concurrency Computat.: Pract. Exper.,2014,26(14):. [17]Stephan E. Korsholm,Hans S?ndergaard,Anders P. Ravn. A real‐time Java tool chain for resource constrained platforms[J]. Concurrency Computat.: Pract. Exper.,2014,26(14):. [18]M. Teresa Higuera‐Toledano,Andy Wellings. Introduction to the Special Issue on Java Technologies for Real‐Time and Embedded Systems: JTRES 2012[J]. Concurrency Computat.: Pract. Exper.,2014,26(14):. [19]Mostafa Mohammadpourfard,Mohammad Ali Doostari,Mohammad Bagher Ghaznavi Ghoushchi,Nafiseh Shakiba. Anew secure Internet voting protocol using Java Card 3 technology and Java information flow concept[J]. Security Comm. Networks,2015,8(2):. [20]Cédric Teyton,Jean‐Rémy Falleri,Marc Palyart,Xavier Blanc. A study of library migrations in Java[J]. J. Softw. Evol. and Proc.,2014,26(11):. [21]Sabela Ramos,Guillermo L. Taboada,Roberto R. Expósito,Juan Touri?o. Nonblocking collectives for scalable Java communications[J]. Concurrency Computat.: Pract. Exper.,2015,27(5):. [22]Dusan Jovanovic,Slobodan Jovanovic. An adaptive e‐learning system for Java programming course, based on Dokeos LE[J]. Comput Appl Eng Educ,2015,23(3):. [23]Yu Lin,Danny Dig. A study and toolkit of CHECK‐THEN‐ACT idioms of Java concurrent collections[J]. Softw. Test. Verif. Reliab.,2015,25(4):. [24]Jonathan Passerat?Palmbach,Claude Mazel,David R. C. Hill. TaskLocalRandom: a statistically sound substitute to pseudorandom number generation in parallel java tasks frameworks[J]. Concurrency Computat.: Pract.Exper.,2015,27(13):. [25]Da Qi,Huaizhong Zhang,Jun Fan,Simon Perkins,Addolorata Pisconti,Deborah M. Simpson,Conrad Bessant,Simon Hubbard,Andrew R. Jones. The mzqLibrary – An open source Java library supporting the HUPO‐PSI quantitative proteomics standard[J]. Proteomics,2015,15(18):. [26]Xiaoyan Zhu,E. James Whitehead,Caitlin Sadowski,Qinbao Song. An analysis of programming language statement frequency in C, C++, and Java source code[J]. Softw. Pract. Exper.,2015,45(11):. [27]Roberto R. Expósito,Guillermo L. Taboada,Sabela Ramos,Juan Touri?o,Ramón Doallo. Low‐latency Java communication devices on RDMA‐enabled networks[J]. Concurrency Computat.: Pract. Exper.,2015,27(17):. [28]V. Serbanescu,K. Azadbakht,F. Boer,C. Nagarajagowda,B. Nobakht. A design pattern for optimizations in data intensive applications using ABS and JAVA 8[J]. Concurrency Computat.: Pract. Exper.,2016,28(2):. [29]E. Tsakalos,J. Christodoulakis,L. Charalambous. The Dose Rate Calculator (DRc) for Luminescence and ESR Dating-a Java Application for Dose Rate and Age Determination[J]. Archaeometry,2016,58(2):. [30]Ronald A. Olsson,Todd Williamson. RJ: a Java package providing JR‐like concurrent programming[J]. Softw. Pract. Exper.,2016,46(5):. java英⽂参考⽂献⼆: [31]Seong‐Won Lee,Soo‐Mook Moon,Seong‐Moo Kim. Flow‐sensitive runtime estimation: an enhanced hot spot detection heuristics for embedded Java just‐in‐time compilers [J]. Softw. Pract. Exper.,2016,46(6):. [32]Davy Landman,Alexander Serebrenik,Eric Bouwers,Jurgen J. Vinju. Empirical analysis of the relationship between CC and SLOC in a large corpus of Java methods and C functions[J]. J. Softw. Evol. and Proc.,2016,28(7):. [33]Renaud Pawlak,Martin Monperrus,Nicolas Petitprez,Carlos Noguera,Lionel Seinturier. SPOON : A library for implementing analyses and transformations of Java source code[J]. Softw. Pract. Exper.,2016,46(9):. [34]Musa Ata?. Open Cezeri Library: A novel java based matrix and computer vision framework[J]. Comput Appl Eng Educ,2016,24(5):. [35]A. Omar Portillo‐Dominguez,Philip Perry,Damien Magoni,Miao Wang,John Murphy. TRINI: an adaptive load balancing strategy based on garbage collection for clustered Java systems[J]. Softw. Pract. Exper.,2016,46(12):. [36]Kim T. Briggs,Baoguo Zhou,Gerhard W. Dueck. Cold object identification in the Java virtual machine[J]. Softw. Pract. Exper.,2017,47(1):. [37]S. Jayaraman,B. Jayaraman,D. Lessa. Compact visualization of Java program execution[J]. Softw. Pract. Exper.,2017,47(2):. [38]Geoffrey Fox. Java Technologies for Real‐Time and Embedded Systems (JTRES2013)[J]. Concurrency Computat.: Pract. Exper.,2017,29(6):. [39]Tórur Biskopst? Str?m,Wolfgang Puffitsch,Martin Schoeberl. Hardware locks for a real‐time Java chip multiprocessor[J]. Concurrency Computat.: Pract. Exper.,2017,29(6):. [40]Serdar Yegulalp. JetBrains' Kotlin JVM language appeals to the Java faithful[J]. ,2016,:. [41]Ortin, Francisco,Conde, Patricia,Fernandez-Lanvin, Daniel,Izquierdo, Raul. The Runtime Performance of invokedynamic: An Evaluation with a Java Library[J]. IEEE Software,2014,31(4):. [42]Johnson, Richard A. JAVA DATABASE CONNECTIVITY USING SQLITE: A TUTORIAL[J]. Allied Academies International Conference. Academy of Information and Management Sciences. Proceedings,2014,18(1):. [43]Trent, Rod. SQL Server Gets PHP Support, Java Support on the Way[J]. SQL Server Pro,2014,:. [44]Foket, C,De Sutter, B,De Bosschere, K. Pushing Java Type Obfuscation to the Limit[J]. IEEE Transactions on Dependable and Secure Computing,2014,11(6):. [45]Parshall, Jon. Rising Sun, Falling Skies: The Disastrous Java Sea Campaign of World War II[J]. United States Naval Institute. Proceedings,2015,141(1):. [46]Brunner, Grant. Java now pollutes your Mac with adware - here's how to uninstall it[J]. ,2015,:. [47]Bell, Jonathan,Melski, Eric,Dattatreya, Mohan,Kaiser, Gail E. Vroom: Faster Build Processes for Java[J]. IEEE Software,2015,32(2):. [48]Chaikalis, T,Chatzigeorgiou, A. Forecasting Java Software Evolution Trends Employing Network Models[J]. IEEE Transactions on Software Engineering,2015,41(6):. [49]Lu, Quan,Liu, Gao,Chen, Jing. Integrating PDF interface into Java application[J]. Library Hi Tech,2014,32(3):. [50]Rashid, Fahmida Y. Oracle fixes critical flaws in Database Server, MySQL, Java[J]. ,2015,:. [51]Rashid, Fahmida Y. Library misuse exposes leading Java platforms to attack[J]. ,2015,:. [52]Rashid, Fahmida Y. Serious bug in widely used Java app library patched[J]. ,2015,:. [53]Odeghero, P,Liu, C,McBurney, PW,McMillan, C. An Eye-Tracking Study of Java Programmers and Application to Source Code Summarization[J]. IEEE Transactions on Software Engineering,2015,41(11):. [54]Greene, Tim. Oracle settles FTC dispute over Java updates[J]. Network World (Online) [55]Rashid, Fahmida Y. FTC ruling against Oracle shows why it's time to dump Java[J]. ,2015,:. [56]Whitwam, Ryan. Google plans to remove Oracle's Java APIs from Android N[J]. ,2015,:. [57]Saher Manaseer,Warif Manasir,Mohammad Alshraideh,Nabil Abu Hashish,Omar Adwan. Automatic Test Data Generation for Java Card Applications Using Genetic Algorithm[J]. Journal of Software Engineering andApplications,2015,8(12):. [58]Paul Venezia. Prepare now for the death of Flash and Java plug-ins[J]. ,2016,:. [59]PW McBurney,C McMillan. Automatic Source Code Summarization of Context for Java Methods[J]. IEEE Transactions on Software Engineering,2016,42(2):. java英⽂参考⽂献三: [61]Serdar Yegulalp,Serdar Yegulalp. Sputnik automates code review for Java projects on GitHub[J].,2016,:. [62]Fahmida Y Rashid,Fahmida Y Rashid. Oracle security includes Java, MySQL, Oracle Database fixes[J]. ,2016,:. [63]H M Chavez,W Shen,R B France,B A Mechling. An Approach to Checking Consistency between UML Class Model and Its Java Implementation[J]. IEEE Transactions on Software Engineering,2016,42(4):. [64]Serdar Yegulalp,Serdar Yegulalp. Unikernel power comes to Java, Node.js, Go, and Python apps[J]. ,2016,:. [65]Yudi Zheng,Stephen Kell,Lubomír Bulej,Haiyang Sun. Comprehensive Multiplatform Dynamic Program Analysis for Java and Android[J]. IEEE Software,2016,33(4):. [66]Fahmida Y Rashid,Fahmida Y Rashid. Oracle's monster security fixes Java, database bugs[J]. ,2016,:. [67]Damian Wolf,Damian Wolf. The top 5 Java 8 features for developers[J]. ,2016,:. [68]Jifeng Xuan,Matias Martinez,Favio DeMarco,Maxime Clément,Sebastian Lamelas Marcote,Thomas Durieux,Daniel LeBerre. Nopol: Automatic Repair of Conditional Statement Bugs in Java Programs[J]. IEEE Transactions on Software Engineering,2017,43(1):. [69]Loo Kang Wee,Hwee Tiang Ning. Vernier caliper and micrometer computer models using Easy Java Simulation and its pedagogical design features-ideas for augmenting learning with real instruments[J]. Physics Education,2014,49(5):. [70]Loo Kang Wee,Tat Leong Lee,Charles Chew,Darren Wong,Samuel Tan. Understanding resonance graphs using Easy Java Simulations (EJS) and why we use EJS[J]. Physics Education,2015,50(2):.【java英⽂参考⽂献汇编】相关⽂章:1.2.3.4.5.6.7.8.。
华中科技大学硕士学位论文摘要随着人工智能的兴起,计算机视觉领域近年来得到快速的发展。
作为计算机视觉领域研究的核心之一,视觉目标跟踪已经广泛应用于许多领域。
由于其应用领域十分广泛,应用场景也变得更加复杂。
尺度变化、遮挡、旋转、复杂背景、低分辨率等因素的存在给视觉目标跟踪带来了的挑战。
基于核相关滤波的KCF(Kernelized Correlation Filters)跟踪算法自问世以来受到了广泛的关注。
本文基于KCF跟踪算法,针对KCF跟踪算法的不足提出了一些改进方法。
针对KCF跟踪算法不能自适应目标尺度的变化,本文提出了SMIKCF(Scale-Adaptive Multiple-Feature Improved-Template-Update KCF)跟踪算法。
SMIKCF在KCF算法的基础上,加入一个尺度估计滤波器;提取样本的HOG((Histogram of Oriented Gradient))特征和CN(Color Name)特征,并把两个特征相融合;利用APCE判据改进位置估计滤波器模型的更新方式。
针对KCF算法在目标发生遮挡时容易出现目标丢失的情况,本文提出了AOKCF(Anti-Occlusion KCF)跟踪算法。
AOKCF在KCF算法的基础上,利用APCE 进行跟踪可靠性检测;加入检测模块,在检测到跟踪结果不可靠时,启动检测模块对目标进行检测,并使用位置滤波器进行目标识别,当识别出目标时,更新目标位置以及位置滤波器,否则直接进入下一帧,继续采用检测模块进行检测。
本文实验数据来源于Visual Tracker Benchmark。
采用OPE(One-Pass Evaluation)的评估方式,在Benchmark的50个测试视频序列上进行实验,来评估算法的性能。
实验表明,本文提出的方法能够有效地解决相关问题,提高跟踪算法的性能。
关键词:目标跟踪核相关滤波尺度自适应遮挡华中科技大学硕士学位论文ABSTRACTWith growing of artificial intelligence, computer vision has been rapid development in recent years. Visual target tracking is one of the most important part of computer vision research, with widely used in many fields. The application situation is becoming more and more complicated due to the wide application fields. The existence of scale changes, occlusion, rotation, complex background, low resolution and other factors has brought higher challenges to visual target tracking. KCF (Kernelized Correlation Filters) tracking algorithm has been widely concerned since it was proposed. Based on the KCF tracking algorithm, this paper proposed improved methods for the shortcoming of KCF tracking algorithm.Firstly, the SMIKCF(Scale-Adaptive Multiple-Feature Improved-Template-Update KCF) tracking algorithm is proposed in this paper to solve the problem that the KCF tracking algorithm can not adapt to the target scale. On the basis of KCF algorithm, SMIKCF adds a scale estimation filter, and combines HOG characteristics and CN characteristics, using the APCE criterion to improve the updating method of the position estimation filter model.Secondly, the AOKCF(Anti-Occlusion KCF) tracking algorithm is proposed to solve the problem of occlusion. AOKCF tracking algorithm is based on KCF tracking algorithm. APCE criterion is used to check the reliability of tracking results. When the tracking result is unreliable, add detection module to detect the target, and then use the position filter to the target recognition. If the target is recognized, then update the target position and the position filter module. Otherwise, go directly to the next frame.The experimental datas are from Visual Tracker Benchmark. Experiments were performed on Benchmark's 50 test video sequences and use the OPE (One-Pass Evaluation) evaluation method to evaluate the performance of the algorithm.华中科技大学硕士学位论文Experimental results show that the proposed method can effectively solve the related problems and improve the performance of the tracking algorithm.Keywords: Object tracking Kernelized correlation filters Scale adaptive Occlusion华中科技大学硕士学位论文目录摘要 (I)Abstract (II)目录 (IV)1 绪论 (1)1.1研究背景与意义 (1)1.2国内外研究概况 (1)1.3论文的主要研究内容 (4)1.4主要组织结构 (4)2 KCF跟踪算法介绍 (6)2.1引言 (6)2.2相关滤波器 (6)2.3核相关滤波器 (9)2.4本章小结 (17)3 改进的尺度自适应KCF跟踪算法研究 (18)3.1引言 (18)3.2尺度估计 (19)3.3多特征融合 (22)3.4模型更新策略改进 (22)3.5算法整体框架 (24)华中科技大学硕士学位论文3.6实验结果与分析 (25)3.7本章小结 (40)4 抗遮挡KCF跟踪算法研究 (41)4.1引言 (41)4.2跟踪可靠性检测 (43)4.3重检测算法研究 (43)4.4算法整体框架 (47)4.5实验结果与分析 (48)4.6本章小结 (60)5 总结与展望 (61)5.1全文总结 (61)5.2课题展望 (62)致谢 (63)参考文献 (64)附录1 攻读学位期间发表论文与参与课题 (70)华中科技大学硕士学位论文1 绪论1.1 研究背景与意义人工智能的飞速发展使得计算机视觉备受关注,作为计算机视觉领域研究的核心之一,视觉目标跟踪已经广泛应用于许多领域[1]。
CVPR 2013 录用论文(目标跟踪部分)/gxf1027/article/details/8650878完整录用论文官方链接:/cvpr13/program.php过段时间CvPaper上面应该会有正文链接今年有关RGB-D摄像机应用和研究的论文渐多起来了。
当然,自己还是比较关心Tracking方面的Papers。
从作者来看,一作大部分为华人,而且有不少在Tracking这个圈子里相当有名的牛,比如Ming-Hsuan Yang,Robert Collins等(中科院到阿大的Xi Li也是非常活跃,从他的论文可以看出深厚的数学功底,另外Chunhua Shen老师团队非常高产)。
此外,从录用论文题目初步判断,Sparse coding (representation)的热度在减退,所以Haibin Ling老师并没有这方面的文章录用,且纯粹的tracking-by-detection几乎不见踪影了。
以下是摘录的tracking方面的录用论文:Oral部分:Structure Preserving Object Tracking. Lu Zhang, Laurens van der MaatenTracking Sports Players with Context-Conditioned Motion Models. Jingchen Liu, Peter Carr, Robert Collins, Yanxi Liu Post部分:Online Object Tracking: A Benchmark. Yi Wu, Jongwoo Lim,Ming-Hsuan YangLearning Compact Binary Codes for Visual Tracking. Xi Li, Chunhua Shen, Anthony Dick, Anton van den HengelPart-based Visual Tracking with Online Latent Structural Learning. Rui Yao, Qinfeng Shi,Chunhua Shen, Yanning Zhang, Anton van den HengelSelf-paced learning for long-term tracking.James Supancic III, Deva Ramanan(long-term的噱头还是很吸引人的,和当年TLD一样,看看是否是工程的思想多一些)Visual Tracking via Locality Sensitive Histograms.Shengfeng He, Qingxiong Yang, Rynson Lau, Jiang Wang,Ming-Hsuan Yang (CityU of HK,使用直方图作为表观在当前研究背景下真是反其道而行之啊)Minimum Uncertainty Gap for Robust Visual Tracking. Junseok Kwon, Kyoung Mu Lee (VTD作者)Least Soft-thresold Squares Tracking. Dong Wang, Huchuan Lu, Ming-Hsuan YangTracking People and Their Objects. Tobias Baumgartner, Dennis Mitzel, Bastian Leibe(这个应该也有应用的背景和前景)(以上不全包括多目标跟踪方面的论文)其它关注的论文:Alternating Decision Forests. Samuel Schulter, Paul Wohlhart,Christian Leistner, Amir Saffari,Peter M. Roth,Horst Bischof(Forest也是近些年的热点之一。
视觉跟踪以连续视频帧为对象,提取目标特征后,根据视频帧序列的上下文信息,利用特征构建目标外观模型,判断候选样本是否为跟踪目标。
跟踪算法按照特征提取方式的不同,通常分为基于传统模型和基于深度学习模型,传统模型通常利用灰度特征、颜色特征、梯度特征以及纹理特征等[1-4]人工特征,单个特征表达单一,跟踪精度不高,融合多个特征之后虽然能够一定程度上提高算法的跟踪精度,但算法对复杂环境下目标出镜头、严重遮挡等情况跟踪效果不佳。
深度学习模型得到的卷积特征具有高层特征更抽象、底层特征更细粒的特点,对目标特征的表达更为丰富[5-6]。
近几年基于深度学习模型的目标跟踪算法在目标丢失判别机制的视觉跟踪算法及应用研究牟清萍1,张莹1,2,张东波1,2,王新杰1,杨知桥11.湘潭大学自动化与电子信息学院,湖南湘潭4111052.机器人视觉感知与控制技术国家工程实验室,长沙410082摘要:当跟踪对象被严重遮挡或者离开相机视野范围时,机器人的跟踪目标往往会丢失。
为了实现准确跟踪,提出了目标丢失判别跟踪YOLO-RTM算法。
该方法通过YOLOv3检测视频第一帧中的目标。
利用实时多域卷积神经网络(Real-Time MDNet,RT-MDNet)跟踪算法预测目标边界框的变化。
计算重叠度,根据重叠度与预设阈值的比较结果决定模型更新方式,当重叠度高于阈值时,采用RT-MDNet更新外观模型,当重叠度低于阈值时,采用YOLOv3重新搜索目标并更新外观模型。
在Turtlebot2机器人上的实验结果表明,提出的算法能满足移动机器人跟踪的可靠性,且有效提高跟踪算法的实用性。
关键词:视觉跟踪;目标丢失判别机制;实时多域卷积神经网络;重叠度;出镜头文献标志码:A中图分类号:TP391doi:10.3778/j.issn.1002-8331.2001-0346Research on Visual Tracking Algorithm and Application of Target Loss Discrimination Mechanism MOU Qingping1,ZHANG Ying1,2,ZHANG Dongbo1,2,WANG Xinjie1,YANG Zhiqiao11.College of Automation and Electronic Information,Xiangtan University,Xiangtan,Hunan411105,China2.National Engineering Laboratory for Robotic Visual Perception and Control Technology,Changsha410082,ChinaAbstract:When the tracked target is severely occluded or leaves the field of camera’s view,the robot tracking target is often lost.In order to track the target robustly,a target loss discrimination tracking algorithm named YOLO-RTM is proposed.This method detects the target in the first frame of video through YOLOv3.Then the Real-Time Multi-Domain convolutional neural Network(Real-Time MDNet,RT-MDNet)tracking algorithm is used to predict the change of the target frame by frame.Finally,the Intersection over Union(IoU)is calculated,and the updating mode of the model is determined according to the comparison result of the IoU and the preset threshold value.If the IoU is higher than the threshold value,RT-MDNet is used to update the appearance model.If the IoU is lower than the threshold value,Yolov3is used to search the target again and update the appearance model.The experimental results on the Turtlebot2robot show that the reliability of mobile robot tracking in general scenes is satisfied by the proposed algorithm,and the practicability of tracking algorithm is improved effectively.Key words:visual tracking;target loss discrimination mechanism;real-time multi-domain convolutional neural networks; Intersection over Union(IoU);leaves the field of camera’s view基金项目:国家自然科学基金(61175075);国家自然科学基金区域创新发展联合基金(U19A2083);湖南省战略性新兴产业科技攻关与重大成果转化项目(2019GK4007);湖南省重点学科项目。
KCF:High-SpeedTrackingwithKernelizedCorrelati。
High-Speed Tracking with Kernelized Correlation Filters 的翻译与分析基于核相关滤波器的⾼速⽬标跟踪⽅法,简称KCF 写在前⾯,之所以对这篇⽂章进⾏精细的阅读,是因为这篇⽂章极其重要,在⽬标跟踪领域⽯破天惊的⼀篇论⽂,后来在此论⽂基础上⼜相继出现了很多基于KCF的⽂章,因此⽂章好⽐作⼤厦的基⽯,深度学习,长短记忆等框架⽹络也可以在KCF上进⾏增添模块,并能够达到较好的效果,因此我将深⼊学习这篇⽂章,并在此与⼤家分享,由于学识有限,难免有些谬误,还请斧正。
High-Speed Tracking with Kernelized Correlation Filters:Abstract---The core component of most modern trackers is discriminative classifier, tasked with distinguishing between the target and the surrounding environment. To cope with natural image changes, this classifier is typically trained with translated and scaled sample patches. Such sets of samples are riddled with redundancies --any overlapping pixels are constrained to be the same. Based on this simple observation, we can diagonalize it with the Discrete Fourier Transform, reducing both storage and computation by several orders of magnitude. Interestingly, for linear regression our formulation is equivalent to a correlation Filter (KCF), that unlike other kernel algorithms has the exact same complexity as its linear counterpart. Building on it ,we also propose a fast multi-channel extension of linear correlation filters ,via a linear kernel ,which we call Dual Correlation Filter(DCF).Both KCF and DCF outperform top-ranking trackers such as Struck of TLD on a 50videos benchmark, despite running at hundreds of frames-per -second , and being implemented in a few lines ofcode(Algorithm1). To encourage further developments, our tracking framework was made open-source.Index Terms——Visual tracking, circulant matrices, discrete Fourier transform, kernel methods, ridge regression, correlation filters.摘要:当前最前沿的跟踪器的核⼼组成是⼀个具有分辨能⼒的分类器,负责将⽬标和其背景环境区分开来。
Online Object Tracking:A BenchmarkYi WuUniversity of California at Merced ywu29@Jongwoo LimHanyang Universityjlim@hanyang.ac.krMing-Hsuan YangUniversity of California at Mercedmhyang@AbstractObject tracking is one of the most important components in numerous applications of computer vision.While much progress has been made in recent years with efforts on shar-ing code and datasets,it is of great importance to develop a library and benchmark to gauge the state of the art.After briefly reviewing recent advances of online object tracking, we carry out large scale experiments with various evalua-tion criteria to understand how these algorithms perform. The test image sequences are annotated with different at-tributes for performance evaluation and analysis.By ana-lyzing quantitative results,we identify effective approaches for robust tracking and provide potential future research di-rections in thisfield.1.IntroductionObject tracking is one of the most important components in a wide range of applications in computer vision,such as surveillance,human computer interaction,and medical imaging[60,12].Given the initialized state(e.g.,position and size)of a target object in a frame of a video,the goal of tracking is to estimate the states of the target in the sub-sequent frames.Although object tracking has been studied for several decades,and much progress has been made in re-cent years[28,16,47,5,40,26,19],it remains a very chal-lenging problem.Numerous factors affect the performance of a tracking algorithm,such as illumination variation,oc-clusion,as well as background clutters,and there exists no single tracking approach that can successfully handle all s-cenarios.Therefore,it is crucial to evaluate the performance of state-of-the-art trackers to demonstrate their strength and weakness and help identify future research directions in this field for designing more robust algorithms.For comprehensive performance evaluation,it is criti-cal to collect a representative dataset.There exist sever-al datasets for visual tracking in the surveillance scenarios, such as the VIVID[13],CA VIAR[21],and PETS databas-es.However,the target objects are usually humans or cars of small size in these surveillance sequences,and the back-ground is usually static.Although some tracking dataset-s[47,5,33]for generic scenes are annotated with bounding box,most of them are not.For sequences without labeled ground truth,it is difficult to evaluate tracking algorithms as the reported results are based on inconsistently annotated object locations.Recently,more tracking source codes have been made publicly available,e.g.,the OAB[22],IVT[47],MIL[5], L1[40],and TLD[31]algorithms,which have been com-monly used for evaluation.However,the input and output formats of most trackers are different and thus it is inconve-nient for large scale performance evaluation.In this work, we build a code library that includes most publicly available trackers and a test dataset with ground-truth annotations to facilitate the evaluation task.Additionally each sequence in the dataset is annotated with attributes that often affect tracking performance,such as occlusion,fast motion,and illumination variation.One common issue in assessing tracking algorithms is that the results are reported based on just a few sequences with different initial conditions or parameters.Thus,the results do not provide the holistic view of these algorithm-s.For fair and comprehensive performance evaluation,we propose to perturb the initial state spatially and temporally from the ground-truth target locations.While the robust-ness to initialization is a well-known problem in thefield, it is seldom addressed in the literature.To the best of our knowledge,this is thefirst comprehensive work to address and analyze the initialization problem of object tracking. We use the precision plots based on location error metric and the success plots based on the overlap metric,to ana-lyze the performance of each algorithm.The contribution of this work is three-fold:Dataset.We build a tracking dataset with50fully annotat-ed sequences to facilitate tracking evaluation.Code library.We integrate most publicly available tracker-s in our code library with uniform input and output formats to facilitate large scale performance evaluation.At present, it includes29tracking algorithms.Robustness evaluation.The initial bounding boxes for tracking are sampled spatially and temporally to evaluate the robustness and characteristics of trackers.Each track-2013 IEEE Conference on Computer Vision and Pattern Recognitioner is extensively evaluated by analyzing more than660,000 bounding box outputs.This work mainly focuses on the online1tracking of sin-gle target.The code library,annotated dataset and all the tracking results are available on the website .2.Related WorkIn this section,we review recent algorithms for object tracking in terms of several main modules:target represen-tation scheme,search mechanism,and model update.In addition,some methods have been proposed that build on combing some trackers or mining context information. Representation Scheme.Object representation is one of major components in any visual tracker and numerous schemes have been presented[35].Since the pioneer-ing work of Lucas and Kanade[37,8],holistic templates (raw intensity values)have been widely used for track-ing[25,39,2].Subsequently,subspace-based tracking ap-proaches[11,47]have been proposed to better account for appearance changes.Furthermore,Mei and Ling[40] proposed a tracking approach based on sparse representa-tion to handle the corrupted appearance and recently it has been further improved[41,57,64,10,55,42].In ad-dition to template,many other visual features have been adopted in tracking algorithms,such as color histogram-s[16],histograms of oriented gradients(HOG)[17,52], covariance region descriptor[53,46,56]and Haar-like fea-tures[54,22].Recently,the discriminative model has been widely adopted in tracking[15,4],where a binary classifier is learned online to discriminate the target from the back-ground.Numerous learning methods have been adapted to the tracking problem,such as SVM[3],structured out-put SVM[26],ranking SVM[7],boosting[4,22],semi-boosting[23]and multi-instance boosting[5].To make trackers more robust to pose variation and partial occlusion, an object can be represented by parts where each one is rep-resented by descriptors or histograms.In[1]several local histograms are used to represent the object in a pre-defined grid structure.Kwon and Lee[32]propose an approach to automatically update the topology of local patches to handle large pose changes.To better handle appearance variation-s,some approaches regarding integration of multiple repre-sentation schemes have recently been proposed[62,51,33]. Search Mechanism.To estimate the state of the target ob-jects,deterministic or stochastic methods have been used. When the tracking problem is posed within an optimiza-tion framework,assuming the objective function is differ-entiable with respect to the motion parameters,gradient descent methods can be used to locate the target efficient-ly[37,16,20,49].However,these objective functions are 1Here,the word online means during tracking only the information of previous few frames is used for inference at any time ually nonlinear and contain many local minima.To allevi-ate this problem,dense sampling methods have been adopt-ed[22,5,26]at the expense of high computational load. On the other hand,stochastic search algorithms such as par-ticlefilters[28,44]have been widely used since they are relatively insensitive to local minima and computationally efficient[47,40,30].Model Update.It is crucial to update the target repre-sentation or model to account for appearance variations. Matthews et al.[39]address the template update problem for the Lucas-Kanade algorithm[37]where the template is updated with the combination of thefixed reference tem-plate extracted from thefirst frame and the result from the most recent frame.Effective update algorithms have also been proposed via online mixture model[29],online boost-ing[22],and incremental subspace update[47].For dis-criminative models,the main issue has been improving the sample collection part to make the online-trained classifier more robust[23,5,31,26].While much progress has been made,it is still difficult to get an adaptive appearance model to avoid drifts.Context and Fusion of Trackers.Context information is also very important for tracking.Recently some approach-es have been proposed by mining auxiliary objects or lo-cal visual information surrounding the target to assist track-ing[59,24,18].The context information is especially help-ful when the target is fully occluded or leaves the image region[24].To improve the tracking performance,some tracker fusion methods have been proposed recently.Sant-ner et al.[48]proposed an approach that combines stat-ic,moderately adaptive and highly adaptive trackers to ac-count for appearance changes.Even multiple trackers[34] or multiple feature sets[61]are maintained and selected in a Bayesian framework to better account for appearance changes.3.Evaluated Algorithms and DatasetsFor fair evaluation,we test the tracking algorithms whose original source or binary codes are publicly avail-able as all implementations inevitably involve technical de-tails and specific parameter settings2.Table1shows the list of the evaluated tracking algorithms.We also evaluate the trackers in the VIVID testbed[13]including the mean shift (MS-V),template matching(TM-V),ratio shift(RS-V)and peak difference(PD-V)methods.In recent years,many benchmark datasets have been de-veloped for various vision problems,such as the Berkeley segmentation[38],FERET face recognition[45]and opti-calflow dataset[9].There exist some datasets for the track-ing in the surveillance scenario,such as the VIVID[13]and CA VIAR[21]datasets.For generic visual tracking,more 2Some source codes[36,58]are obtained from direct contact,and some methods are implemented on our own[44,16].Method Representation Search MU Code FPSCPF [44]L,IH PF N C109LOT[43]L,color PF Y M0.70IVT[47]H,PCA,GM PF Y MC33.4ASLA[30]L,SR,GM PF Y MC8.5SCM[65]L,SR,GM+DM PF Y MC0.51L1APG[10]H,SR,GM PF Y MC 2.0MTT[64]H,SR,GM PF Y M 1.0VTD[33]H,SPCA,GM MCMC Y MC-E 5.7VTS[34]L,SPCA,GM MCMC Y MC-E 5.7LSK[36]L,SR,GM LOS Y M-E 5.5ORIA[58]H,T,GM LOS Y M9.0DFT[49]L,T LOS Y M13.2KMS[16]H,IH LOS N C3,159SMS[14]H,IH LOS N C19.2VR-V[15]H,color LOS Y MC109Frag[1]L,IH DS N C 6.3OAB[22]H,Haar,DM DS Y C22.4SemiT[23]H,Haar,DM DS Y C11.2BSBT[50]H,Haar,DM DS Y C7.0MIL[5]H,Haar,DM DS Y C38.1CT[63]H,Haar,DM DS Y MC64.4TLD[31]L,BP,DM DS Y MC28.1Struck[26]H,Haar,DM DS Y C20.2CSK[27]H,T,DM DS Y M362CXT[18]H,BP,DM DS Y C15.3Table1.Evaluated tracking algorithms(MU:model update,FP-S:frames per second).For representation schemes,L:local,H:holistic,T:template,IH:intensity histogram,BP:binary pattern,PCA:principal component analysis,SPCA:sparse PCA,SR:s-parse representation,DM:discriminative model,GM:generativemodel.For search mechanism,PF:particlefilter,MCMC:MarkovChain Monte Carlo,LOS:local optimum search,DS:dense sam-pling search.For the model update,N:No,Y:Yes.In the Codecolumn,M:Matlab,C:C/C++,MC:Mixture of Matlab and C/C++,suffix E:executable binary code.sequences have been used for evaluation[47,5].However,most sequences do not have the ground truth annotation-s,and the quantitative evaluation results may be generatedwith different initial conditions.To facilitate fair perfor-mance evaluation,we have collected and annotated mostcommonly used tracking sequences.Figure1shows thefirstframe of each sequence where the target object is initializedwith a bounding box.Attributes of a test sequence.Evaluating trackers is dif-ficult because many factors can affect the tracking perfor-mance.For better evaluation and analysis of the strengthand weakness of tracking approaches,we propose to catego-rize the sequences by annotating them with the11attributesshown in Table2.The attribute distribution in our dataset is shown in Fig-ure2(a).Some attributes occur more frequently,e.g.,OPRand IPR,than others.It also shows that one sequence isoften annotated with several attributes.Aside from summa-rizing the performance on the whole dataset,we also con-struct several subsets corresponding to attributes to reportspecific challenging conditions.For example,the OCC sub-set contains29sequences which can be used to analyze theAttr DescriptionIV Illumination Variation-the illumination in the target region issignificantly changed.SV Scale Variation-the ratio of the bounding boxes of thefirstframe and the current frame is out of the range[1/t s,t s],t s>1(t s=2).OCC Occlusion-the target is partially or fully occluded.DEF Deformation-non-rigid object deformation.MB Motion Blur-the target region is blurred due to the motion oftarget or camera.FM Fast Motion-the motion of the ground truth is larger than t mpixels(t m=20).IPR In-Plane Rotation-the target rotates in the image plane.OPR Out-of-Plane Rotation-the target rotates out of the imageplane.OV Out-of-View-some portion of the target leaves the view.BC Background Clutters-the background near the target has thesimilar color or texture as the target.LR Low Resolution-the number of pixels inside the ground-truthbounding box is less than t r(t r=400).Table2.List of the attributes annotated to test sequences.Thethreshold values used in this work are also shown.(a)(b)Figure2.(a)Attribute distribution of the entire testset,and(b)thedistribution of the sequences with occlusion(OCC)attribute.performance of trackers to handle occlusion.The attributedistributions in OCC subset is shown in Figure2(b)and oth-ers are available in the supplemental material.4.Evaluation MethodologyIn this work,we use the precision and success rate forquantitative analysis.In addition,we evaluate the robust-ness of tracking algorithms in two aspects.Precision plot.One widely used evaluation metric on track-ing precision is the center location error,which is definedas the average Euclidean distance between the center loca-tions of the tracked targets and the manually labeled groundtruths.Then the average center location error over all theframes of one sequence is used to summarize the overallperformance for that sequence.However,when the trackerloses the target,the output location can be random and theaverage error value may not measure the tracking perfor-mance correctly[6].Recently the precision plot[6,27]hasbeen adopted to measure the overall tracking performance.It shows the percentage of frames whose estimated locationis within the given threshold distance of the ground truth.As the representative precision score for each tracker weuse the score for the threshold=20pixels[6].Success plot.Another evaluation metric is the boundingbox overlap.Given the tracked bounding box r t and theFigure 1.Tracking sequences for evaluation.The first frame with the bounding box of the target object is shown for each sequence.The sequences are ordered based on our ranking results (See supplementary material):the ones on the top left are more difficult for tracking than the ones on the bottom right.Note that we annotated two targets for the jogging sequence.ground truth bounding box r a ,the overlap score is definedas S =|r t r a ||r t r a |,where and represent the intersec-tion and union of two regions,respectively,and |·|denotes the number of pixels in the region.To measure the perfor-mance on a sequence of frames,we count the number of successful frames whose overlap S is larger than the given threshold t o .The success plot shows the ratios of success-ful frames at the thresholds varied from 0to ing one success rate value at a specific threshold (e.g.t o =0.5)for tracker evaluation may not be fair or representative.Instead we use the area under curve (AUC)of each success plot to rank the tracking algorithms.Robustness Evaluation.The conventional way to evaluate trackers is to run them throughout a test sequence with ini-tialization from the ground truth position in the first frame and report the average precision or success rate.We re-fer this as one-pass evaluation (OPE ).However a tracker may be sensitive to the initialization,and its performance with different initialization at a different start frame may become much worse or better.Therefore,we propose two ways to analyze a tracker’s robustness to initialization,by perturbing the initialization temporally (i.e.,start at differ-ent frames)and spatially (i.e.,start by different bounding boxes).These tests are referred as temporal robustness e-valuation (TRE )and spatial robustness evaluation (SRE )respectively.The proposed test scenarios happen a lot in the real-world applications as a tracker is often initialized by an ob-ject detector,which is likely to introduce initialization er-rors in terms of position and scale.In addition,an object detector may be used to re-initialize a tracker at differen-t time instances.By investigating a tracker’s characteristic in the robustness evaluation,more thorough understandingand analysis of the tracking algorithm can be carried out.Temporal Robustness Evaluation.Given one initial frame together with the ground-truth bounding box of target,one tracker is initialized and runs to the end of the sequence,i.e.,one segment of the entire sequence.The tracker is evaluat-ed on each segment,and the overall statistics are tallied.Spatial Robustness Evaluation.We sample the initial bounding box in the first frame by shifting or scaling the ground truth.Here,we use 8spatial shifts including 4cen-ter shifts and 4corner shifts,and 4scale variations (supple-ment).The amount for shift is 10%of target size,and the scale ratio varys among 0.8,0.9,1.1and 1.2to the ground truth.Thus,we evaluate each tracker 12times for SRE.5.Evaluation ResultsFor each tracker,the default parameters with the source code are used in all evaluations.Table 1lists the average FPS of each tracker in OPE running on a PC with Intel i73770CPU (3.4GHz).More detailed speed statistics,such as minimum and maximum,are available in the supplement.For OPE,each tracker is tested on more than 29,000frames.For SRE,each tracker is evaluated 12times on each sequence,where more than 350,000bounding box results are generated.For TRE,each sequence is partitioned into 20segments and thus each tracker is performed on around 310,000frames.To the best of our knowledge,this is the largest scale performance evaluation of visual tracking.We report the most important findings in this manuscript and more details and figures can be found in the supplement.5.1.Overall PerformanceThe overall performance for all the trackers is summa-rized by the success and precision plots as shown in Fig-Figure3.Plots of OPE,SRE,and TRE.The performance score for each tracker is shown in the legend.For eachfigure,the top10trackers are presented for clarity and complete plots are in the supplementary material(best viewed on high-resolution display).ure3where only the top10algorithms are presented for clarity and the complete plots are displayed in the supple-mentary material.For success plots,we use AUC scores to summarize and rank the trackers,while for precision plots we use the results at error threshold of20for ranking.In the precision plots,the rankings of some trackers are slight-ly different from the rankings in the success plots in that they are based on different metrics which measure different characteristics of trackers.Because the AUC score of suc-cess plot measures the overall performance which is more accurate than the score at one threshold of the plot,in the following we mainly analyze the rankings based on success plots but use the precision plots as auxiliary.The average TRE performance is higher than that of OPE in that the number of frames decreases from thefirst to last segment of TRE.As the trackers tend to perform well in shorter sequences,the average of all the results in TRE tend to be higher.On the other hand,the average performance of SRE is lower than that of OPE.The initialization errors tend to cause trackers to update with imprecise appearance information,thereby causing gradual drifts.In the success plots,the top ranked tracker SCM in OPE outperforms Struck by2.6%but is1.9%below Struck in SRE.The results also show that OPE is not the best perfor-mance indicator as the OPE is one trial of SRE or TRE.The ranking of TLD in TRE is lower than OPE and SRE.This is because TLD performs well in long sequences with a re-detection module while there are numerous short segments in TRE.The success plots of Struck in TRE and SRE show that the success rate of Struck is higher than SCM and AL-SA when the overlap threshold is small,but less than SCM and ALSA when the overlap threshold is large.This is be-cause Struck only estimates the location of target and does not handle scale variation.Sparse representations are used in SCM,ASLA,LSK, MTT and L1APG.These trackers perform well in SRE and TRE,which suggests sparse representations are effective models to account for appearance change(e.g.,occlusion). We note that SCM,ASLA and LSK outperform MTT and L1APG.The results suggest that local sparse representa-tions are more effective than the ones with holistic sparse templates.The AUC score of ASLA deceases less than the other top5trackers from OPE to SRE and the rank-ing of ASLA also increases.It indicates the alignment-pooling technique adopted by ASLA is more robust to mis-alignments and background clutters.Among the top10trackers,CSK has the highest speed where the proposed circulant structure plays a key role. The VTD and VTS methods adopt mixture models to im-prove the tracking pared with other high-er ranked trackers,the performance bottleneck of them can be attributed to their adopted representation based on sparseprincipal component analysis,where the holistic templates are used.Due to the space limitation,the plots of SRE are presented for analysis in the following sections,and more results are included in the supplement.5.2.Attribute-based Performance AnalysisBy annotating the attributes of each sequence,we con-struct subsets with different dominant attributes which facil-itates analyzing the performance of trackers for each chal-lenging factor.Due to space limitations,we only illustrate and analyze the success plots and precision plots of SRE for attributes OCC,SV,and FM as shown in Figure4,and more results are presented in the supplementary material.When an object moves fast,dense sampling based track-ers(e.g.,Struck,TLD and CXT)perform much better than others.One reason is that the search ranges are large and the discriminative models are able to discriminate the tar-gets from the background clutters.However,the stochastic search based trackers with high overall performance(e.g., SCM and ASLA)do not perform well in this subset due to the poor dynamic models.If these parameters are set to large values,more particles are required to make the tracker stable.These trackers can be further improved with dynam-ic models with more effective particlefilters.On the OCC subset,the Struck,SCM,TLD,LSK and ASLA methods outperform others.The results suggest that structured learning and local sparse representations are effective in dealing with occlusions.On the SV subset, ASLA,SCM and Struck perform best.The results show that trackers with affine motion models(e.g.,ASLA and SCM) often handle scale variation better than others that are de-signed to account for only translational motion with a few exceptions such as Struck.5.3.Initialization with Different ScaleIt has been known that trackers are often sensitive to ini-tialization variations.Figure5and Figure6show the sum-marized tracking performance with initialization at differ-ent scales.When computing the overlap score,we rescale the tracking results so that the performance summary could be comparable with the original scale,i.e.,the plots of OPE in Figure3.Figure6illustrates the average performance of all trackers for each scale which shows the performance often decreases significantly when the scale factor is large (e.g.,×1.2)as many background pixels are inevitably in-cluded in the initial representations.The performance of TLD,CXT,DFT and LOT decreases with the increase of initialization scale.This indicates these trackers are more sensitive to background clutters.Some trackers perform better when the scale factor is smaller,such as L1APG,MT-T,LOT and CPF.One reason for this in the case of L1APG and MTT is that the templates have to be warped tofit the size of the usually smaller canonical template so that if the initial template is small,more appearance details will be kept in the model.On the other hand,some trackers per-form well or even better when the initial bounding box is enlarged,such as Struck,OAB,SemiT,and BSBT.This in-dicates that the Haar-like features are somewhat robust to background clutters due to the summation operations when computing features.Overall,Struck is less sensitive to scale variation than other well-performing methods.6.Concluding RemarksIn this paper,we carry out large scale experiments to evaluate the performance of recent online tracking algo-rithms.Based on our evaluation results and observations, we highlight some tracking components which are essen-tial for improving tracking performance.First,background information is critical for effective tracking.It can be ex-ploited by using advanced learning techniques to encode the background information in the discriminative model im-plicitly(e.g.,Struck),or serving as the tracking contex-t explicitly(e.g.,CXT).Second,local models are impor-tant for tracking as shown in the performance improvement of local sparse representation(e.g.,ASLA and SCM)com-pared with the holistic sparse representation(e.g.,MTT and L1APG).They are particularly useful when the appearance of target is partially changed,such as partial occlusion or deformation.Third,motion model or dynamic model is cru-cial for object tracking,especially when the motion of target is large or abrupt.However,most of our evaluated tracker-s do not focus on this component.Good location predic-tion based on the dynamic model could reduce the search range and thus improve the tracking efficiency and robust-ness.Improving these components will further advance the state of the art of online object tracking.The evaluation results show that significant progress in thefield of object tracking has been made in the last decade. We propose and demonstrate evaluation metrics for in-depth analysis of tracking algorithms from several perspectives. This large scale performance evaluation facilitates better understanding of the state-of-the-art online object tracking approaches,and provides a platform for gauging new algo-rithms.Our ongoing work focuses on extending the dataset and code library to include more fully annotated sequences and trackers.Acknowledgment.We thank the reviewers for valuable comments and suggestions.The work is supported partly by NSF CAREER Grant#1149783and NSF IIS Grant#1152576.Wu is also with Nanjing University of Information Science and Technology,China and supported partly by NSFC Grant#61005027. References[1] A.Adam,E.Rivlin,and I.Shimshoni.Robust Fragments-basedTracking using the Integral Histogram.In CVPR,2006.[2]N.Alt,S.Hinterstoisser,and N.Navab.Rapid Selection of ReliableTemplates for Visual Tracking.In CVPR,2010.[3]S.Avidan.Support Vector Tracking.PAMI,26(8):1064–1072,2004.[4]S.Avidan.Ensemble Tracking.PAMI,29(2):261–271,2008.。