DiTracker: Repurposing
Video Diffusion Transformers for Robust Point Tracking

arXiv 2025

1KAIST AI 2Google DeepMind
†: Co-corresponding author
TL;DR

DiTracker repurposes video DiTs for point tracking with softmax-based matching, LoRA adaptation, and cost fusion, achieving stronger robustness and faster convergence.


DiTracker teaser graph

DiTracker leverages pre-trained video Diffusion Transformer (DiT) features to outperform state-of-the-art methods such as CoTracker3 in challenging real-world scenarios involving complex motion and frequent occlusions, while achieving comparable final performance with 10x faster convergence, substantially reducing training cost. These results demonstrate that pre-trained video DiT features constitute an effective and efficient foundation for robust point tracking

Abstract

Point tracking aims to localize corresponding points across video frames, serving as a fundamental task for 4D reconstruction, robotics, and video editing. Existing methods commonly rely on shallow convolutional backbones such as ResNet that process frames independently, lacking temporal coherence and producing unreliable matching costs under challenging conditions. Through systematic analysis, we find that video Diffusion Transformers (DiTs), pre-trained on large-scale real-world videos with spatio-temporal attention, inherently exhibit strong point tracking capability and robustly handle dynamic motions and frequent occlusions. We propose DiTracker, which adapts video DiTs through: (1) query-key attention matching, (2) lightweight LoRA tuning, and (3) cost fusion with a ResNet backbone. Despite training with 8x smaller batch size, DiTracker achieves state-of-the-art performance on challenging ITTO benchmark and matches or outperforms state-of-the-art models on TAP-Vid benchmarks. Our work validates video DiT features as an effective and efficient foundation for point tracking.

Analysis

To investigate whether pre-trained video DiT features can address key limitations in point tracking, our analysis compares multiple pre-trained video DiT models (HunyuanVideo, CogVideoX-2B, CogVideoX-5B, Wan-14B) with the supervised ResNet backbone from CoTracker3 and other vision foundation models (DINOv2, DINOv3, V-JEPA2).

Best 2nd 3rd
Method < δ0 < δ1 < δ2 < δ3 < δ4 δxavg
ResNet (CoTracker3) 10.5 34.0 49.7 57.7 66.6 43.7
V-JEPA2 2.810.530.555.469.933.8
DINOv2-B/14 2.810.935.667.283.540.0
DINOv3-B/16 3.011.837.768.284.741.1
HunyuanVideo 4.418.244.870.182.844.1
CogVideoX-2B 4.819.449.273.686.346.3
CogVideoX-5B 5.220.550.773.984.346.9
WAN-14B 12.431.956.772.182.751.2

Zero-shot point tracking performance comparison on TAP-Vid-DAVIS benchmark. Video DiTs consistently provide superior initial matching despite having no tracking-specific training.

Best 2nd 3rd
Method Motion Blur Track Motion Reappearance Frequency
12345 [0%, 0.5%)[0.5%, 1.5%)[1.5%, 5%) [0, 1)[1, 3)[3, 100)
ResNet (CoTracker3) 42.440.036.231.027.9 24.716.79.7 22.717.110.6
V-JEPA2 29.027.324.320.319.0 21.823.217.4 24.420.917.0
DINOv2 39.037.634.732.429.9 24.323.519.3 23.923.619.9
DINOv3 40.038.536.534.031.8 30.124.420.8 29.125.820.1
HunyuanVideo 41.740.638.936.434.5 42.730.716.9 40.628.719.4
CogVideoX-2B 45.043.641.639.036.7 45.433.719.2 43.131.721.5
CogVideoX-5B 45.544.242.239.537.7 44.937.422.2 44.534.123.9
WAN-14B 50.749.647.644.642.5 46.335.321.2 43.434.322.9

Analysis on challenging real-world scenarios from ITTO-MOSE benchmark: (1) motion blur, (2) dynamic motion and (3) frequent occlusions. The consistent superiority of all video DiTs validates that large-scale pretraining with full 3D attention provides fundamentally stronger motion priors for challenging correspondence tasks.

DiTracker

DiTracker method overview

Overall Architecture of DiTracker. For long video sequences, input frames are divided into \( N \) temporal chunks with the global first frame prepended. Individual video frames are encoded via a VAE and processed by a video DiT to extract query features \( q_i \) and key features \( k_j \).

\[ \mathcal{C}^{\mathrm{DiT}}_{i,j} = \mathrm{Softmax}\!\left( \frac{q_i k_j^\top}{\sqrt{d}} \right) \]

The DiT local cost is fused with the ResNet local cost \( \mathcal{C}^{\mathrm{ResNet}}_{i,j} \). Finally, a tracking head refines trajectories over \( T \) iterations, updating displacement \( \Delta P \), visibility \( \Delta V \), and confidence \( \Delta C \).

Quantitative Results

Qualitative Results

ITTO-MOSE

TAP-Vid-DAVIS

TAP-Vid-DAVIS with Corruptions

Citation


      @misc{son2025repurposingvideodiffusiontransformers,
            title={Repurposing Video Diffusion Transformers for Robust Point Tracking}, 
            author={Soowon Son and Honggyu An and Chaehyun Kim and Hyunah Ko and Jisu Nam and Dahyun Chung and Siyoon Jin and Jung Yi and Jaewon Min and Junhwa Hur and Seungryong Kim},
            year={2025},
            eprint={2512.20606},
            archivePrefix={arXiv},
            primaryClass={cs.CV},
            url={https://arxiv.org/abs/2512.20606}, 
      }