TY - GEN
T1 - Pose-configurable generic tracking of elongated objects
AU - Wesierski, Daniel
AU - Horain, Patrick
PY - 2013/1/1
Y1 - 2013/1/1
N2 - Elongated objects have various shapes and can shift, rotate, change scale, and be rigid or deform by flexing, articulating, and vibrating, with examples as varied as a glass bottle, a robotic arm, a surgical suture, a finger pair, a tram, and a guitar string. This generally makes tracking of poses of elongated objects very challenging. We describe a unified, configurable framework for tracking the pose of elongated objects, which move in the image plane and extend over the image region. Our method strives for simplicity, versatility, and efficiency. The object is decomposed into a chained assembly of segments of multiple parts that are arranged under a hierarchy of tailored spatio-temporal constraints. In this hierarchy, segments can rescale independently while their elasticity is controlled with global orientations and local distances. While the trend in tracking is to design complex, structure-free algorithms that update object appearance on-line, we show that our tracker, with the novel but remarkably simple, structured organization of parts with constant appearance, reaches or improves state-of-the-art performance. Most importantly, our model can be easily configured to track exact pose of arbitrary, elongated objects in the image plane. The tracker can run up to 100 fps on a desktop PC, yet the computation time scales linearly with the number of object parts. To our knowledge, this is the first approach to generic tracking of elongated objects.
AB - Elongated objects have various shapes and can shift, rotate, change scale, and be rigid or deform by flexing, articulating, and vibrating, with examples as varied as a glass bottle, a robotic arm, a surgical suture, a finger pair, a tram, and a guitar string. This generally makes tracking of poses of elongated objects very challenging. We describe a unified, configurable framework for tracking the pose of elongated objects, which move in the image plane and extend over the image region. Our method strives for simplicity, versatility, and efficiency. The object is decomposed into a chained assembly of segments of multiple parts that are arranged under a hierarchy of tailored spatio-temporal constraints. In this hierarchy, segments can rescale independently while their elasticity is controlled with global orientations and local distances. While the trend in tracking is to design complex, structure-free algorithms that update object appearance on-line, we show that our tracker, with the novel but remarkably simple, structured organization of parts with constant appearance, reaches or improves state-of-the-art performance. Most importantly, our model can be easily configured to track exact pose of arbitrary, elongated objects in the image plane. The tracker can run up to 100 fps on a desktop PC, yet the computation time scales linearly with the number of object parts. To our knowledge, this is the first approach to generic tracking of elongated objects.
KW - configurable tracking
KW - generic tracking
KW - tracking elongated objects
KW - tracking longitudinal objects
KW - visual object tracking
UR - https://www.scopus.com/pages/publications/84898800398
U2 - 10.1109/ICCV.2013.363
DO - 10.1109/ICCV.2013.363
M3 - Conference contribution
AN - SCOPUS:84898800398
SN - 9781479928392
T3 - Proceedings of the IEEE International Conference on Computer Vision
SP - 2920
EP - 2927
BT - Proceedings - 2013 IEEE International Conference on Computer Vision, ICCV 2013
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2013 14th IEEE International Conference on Computer Vision, ICCV 2013
Y2 - 1 December 2013 through 8 December 2013
ER -