TY - GEN
T1 - Single object tracking using offline trained deep regression networks
AU - Mocanu, Bogdan
AU - Tapu, Ruxandra
AU - Zaharia, Titus
N1 - Publisher Copyright:
© 2017 IEEE.
PY - 2017/7/2
Y1 - 2017/7/2
N2 - In this paper we introduce a novel single object tracker based on two convolutional neural networks (CNNs) trained offline using data from large videos repositories. The key principle consists of alternating between tracking using motion information and adjusting the predicted location based on visual similarity. First, we construct a deep regression network architecture able to learn generic relations between the object appearance models and its associated motion patterns. Then, based on visual similarity constraints, the objects bounding box position, size and shape are continuously updated in order to maximize a patch similarity function designed using CNN. Finally, a multi-resolution fusion between the outputs of the two CNNs is performed for accurate object localization. The experimental evaluation performed on challenging datasets, proposed in the visual object tracking (VOT) international contest, validates the proposed method when compared with state-of-the-art systems. In terms of computational speed our tracker runs at 20fps.
AB - In this paper we introduce a novel single object tracker based on two convolutional neural networks (CNNs) trained offline using data from large videos repositories. The key principle consists of alternating between tracking using motion information and adjusting the predicted location based on visual similarity. First, we construct a deep regression network architecture able to learn generic relations between the object appearance models and its associated motion patterns. Then, based on visual similarity constraints, the objects bounding box position, size and shape are continuously updated in order to maximize a patch similarity function designed using CNN. Finally, a multi-resolution fusion between the outputs of the two CNNs is performed for accurate object localization. The experimental evaluation performed on challenging datasets, proposed in the visual object tracking (VOT) international contest, validates the proposed method when compared with state-of-the-art systems. In terms of computational speed our tracker runs at 20fps.
KW - Single object tracking
KW - convolutional neural networks
KW - object appearance model
KW - occlusion detection
UR - https://www.scopus.com/pages/publications/85050666440
U2 - 10.1109/IPTA.2017.8310091
DO - 10.1109/IPTA.2017.8310091
M3 - Conference contribution
AN - SCOPUS:85050666440
T3 - Proceedings of the 7th International Conference on Image Processing Theory, Tools and Applications, IPTA 2017
SP - 1
EP - 6
BT - Proceedings of the 7th International Conference on Image Processing Theory, Tools and Applications, IPTA 2017
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 7th International Conference on Image Processing Theory, Tools and Applications, IPTA 2017
Y2 - 28 November 2017 through 1 December 2017
ER -