文章目录
- 1. ARCF: "Learning Aberrance Repressed Correlation Filters for Real-Time UAV Tracking"
- 2. MLT: "Deep Meta Learning for Real-Time Target-Aware Visual Tracking"
- 3. : "Joint Monocular 3D Vehicle Detection and Tracking"
- 4. : "`Skimming-Perusal' Tracking: A Framework for Real-Time and Robust Long-term Tracking"
- 5. DiMP: "Learning Discriminative Model Prediction for Tracking"
- 6. : "Joint Group Feature Selection and Discriminative Filter Learning for Robust Visual Object Tracking"
- 7. GradNet: "GradNet: Gradient-Guided Network for Visual Object Tracking"
- 8. : "Bridging the Gap Between Detection and Tracking: A Unified Approach"
- 9. : "Physical Adversarial Textures That Fool Visual Object Tracking"
- 10. CDTB : "CDTB: A Color and Depth Visual Object Tracking Dataset and Benchmark"
1. ARCF: “Learning Aberrance Repressed Correlation Filters for Real-Time UAV Tracking”
[在线阅读]
Figure 1. Comparison between background-aware correlation filter (BACF) and the proposed ARCF tracker. The central figure is to demonstrate the differences between previous response map and current response map on group1 1 from UAV123@10fps. Sudden changes of response maps indicate aberrances. When aberrances take place, BACF is tend to lose track of the object while the proposed ARCF can repress aberrances so that this kind of drifting can be avoided.
Figure 2. Main work-flow of the proposed ARCF tracker. It learns both positive sample (green samples) of the object and negative samples (red samples) extracted from the background and the response map restriction is integrated in the learning process so that aberrances in response maps can be repressed. [ ψ p , q ] \left[\psi_{p, q}\right] [ψp,q] serves to shift the generated response map so that the peak position in the previous frame is the same as that of the current frame and thus the position of the detected object will not affect the restriction.
2. MLT: “Deep Meta Learning for Real-Time Target-Aware Visual Tracking”
[在线阅读]
Figure 1: Motivation of the proposed visual tracker. Our framework incorporates a meta-learner network along with a matching network. The meta-learner network receives meta information from the matching network and provides the matching network with the adaptive target-specific feature space needed for robust matching and tracking.
Figure 2: Overview of proposed visual tracking framework. The matching network provides the meta-learner network with meta-information in the form of loss gradients obtained using the training samples. Then the meta-learner network provides the matching network with target-specific information in the form of convolutional kernels and channel-wise attention.
Figure 3: Training scheme of meta-learner network. The meta-learner network uses loss gradients δ δ δ in (2) as meta information, derived from the matching network, which explains its own status in the current feature space [35]. Then, the function g ( ⋅ ) g(·) g