Title :
Online multi-modal task-driven dictionary learning and robust joint sparse representation for visual tracking
Author :
Ali Taalimi;Hairong Qi;Rahman Khorsandi
Author_Institution :
Electrical Engineering and Computer Science, The University of Tennessee, Knoxville, 1520 Middle Drive, United States
Abstract :
Robust visual tracking is a challenging problem due to pose variance, occlusion and cluttered backgrounds. No single feature can be robust to all possible scenarios in a video sequence. However, exploiting multiple features has demonstrated its effectiveness in overcoming challenging situations in visual tracking. We propose a new framework for multi-modal fusion at both the feature level and decision level by training a reconstructive and discriminative dictionary and classifier for each modality simultaneously with the additional constraint of label consistency across different modalities. In addition, a joint decision measure is designed based on both reconstruction and classification error to adaptively adjust the weights of different features such that unreliable features can be removed from tracking. The proposed tracking scheme is referred to as the label-consistent and fusion-based joint sparse coding (LC-FJSC). Extensive experiments on publicly available videos demonstrate that LC-FJSC outperforms state-of-the-art trackers.
Keywords :
"Dictionaries","Joints","Visualization","Target tracking","Yttrium","Encoding","Training"
Conference_Titel :
Advanced Video and Signal Based Surveillance (AVSS), 2015 12th IEEE International Conference on
DOI :
10.1109/AVSS.2015.7301775