计算机科学
人工智能
BitTorrent跟踪器
稳健性(进化)
计算机视觉
动画
视频跟踪
目标检测
跳跃式监视
跟踪(教育)
深度学习
任务(项目管理)
机器学习
对象(语法)
眼动
模式识别(心理学)
计算机图形学(图像)
管理
化学
经济
心理学
基因
生物化学
教育学
作者
Zhen He,Jian Li,Daxue Liu,Hangen He,David Barber
标识
DOI:10.1109/cvpr.2019.00141
摘要
Online Multi-Object Tracking (MOT) from videos is a challenging computer vision task which has been extensively studied for decades. Most of the existing MOT algorithms are based on the Tracking-by-Detection (TBD) paradigm combined with popular machine learning approaches which largely reduce the human effort to tune algorithm parameters. However, the commonly used supervised learning approaches require the labeled data (e.g., bounding boxes), which is expensive for videos. Also, the TBD framework is usually suboptimal since it is not end-to-end, i.e., it considers the task as detection and tracking, but not jointly. To achieve both label-free and end-to-end learning of MOT, we propose a Tracking-by-Animation framework, where a differentiable neural model first tracks objects from input frames and then animates these objects into reconstructed frames. Learning is then driven by the reconstruction error through backpropagation. We further propose a Reprioritized Attentive Tracking to improve the robustness of data association. Experiments conducted on both synthetic and real video datasets show the potential of the proposed model. Our project page is publicly available at: https://github.com/zhen-he/tracking-by-animation
科研通智能强力驱动
Strongly Powered by AbleSci AI