Robust Visual Tracking via Exclusive Context Modeling

Tianzhu Zhang, Bernard Ghanem, Si Liu, Changsheng Xu, Narendra Ahuja

Research output: Contribution to journalArticlepeer-review

48 Scopus citations

Abstract

In this paper, we formulate particle filter-based object tracking as an exclusive sparse learning problem that exploits contextual information. To achieve this goal, we propose the context-aware exclusive sparse tracker (CEST) to model particle appearances as linear combinations of dictionary templates that are updated dynamically. Learning the representation of each particle is formulated as an exclusive sparse representation problem, where the overall dictionary is composed of multiple {group} dictionaries that can contain contextual information. With context, CEST is less prone to tracker drift. Interestingly, we show that the popular L₁ tracker [1] is a special case of our CEST formulation. The proposed learning problem is efficiently solved using an accelerated proximal gradient method that yields a sequence of closed form updates. To make the tracker much faster, we reduce the number of learning problems to be solved by using the dual problem to quickly and systematically rank and prune particles in each frame. We test our CEST tracker on challenging benchmark sequences that involve heavy occlusion, drastic illumination changes, and large pose variations. Experimental results show that CEST consistently outperforms state-of-the-art trackers.
Original languageEnglish (US)
Pages (from-to)51-63
Number of pages13
JournalIEEE Transactions on Cybernetics
Volume46
Issue number1
DOIs
StatePublished - Feb 9 2015

Fingerprint

Dive into the research topics of 'Robust Visual Tracking via Exclusive Context Modeling'. Together they form a unique fingerprint.

Cite this