TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/TrackSSM: A General Motion Predictor by State-Space Model

TrackSSM: A General Motion Predictor by State-Space Model

Bin Hu, Run Luo, Zelin Liu, Cheng Wang, Wenyu Liu

2024-08-31Multi-Object TrackingObject TrackingMultiple Object Tracking
PaperPDFCode(official)

Abstract

Temporal motion modeling has always been a key component in multiple object tracking (MOT) which can ensure smooth trajectory movement and provide accurate positional information to enhance association precision. However, current motion models struggle to be both efficient and effective across different application scenarios. To this end, we propose TrackSSM inspired by the recently popular state space models (SSM), a unified encoder-decoder motion framework that uses data-dependent state space model to perform temporal motion of trajectories. Specifically, we propose Flow-SSM, a module that utilizes the position and motion information from historical trajectories to guide the temporal state transition of object bounding boxes. Based on Flow-SSM, we design a flow decoder. It is composed of a cascaded motion decoding module employing Flow-SSM, which can use the encoded flow information to complete the temporal position prediction of trajectories. Additionally, we propose a Step-by-Step Linear (S$^2$L) training strategy. By performing linear interpolation between the positions of the object in the previous frame and the current frame, we construct the pseudo labels of step-by-step linear training, ensuring that the trajectory flow information can better guide the object bounding box in completing temporal transitions. TrackSSM utilizes a simple Mamba-Block to build a motion encoder for historical trajectories, forming a temporal motion model with an encoder-decoder structure in conjunction with the flow decoder. TrackSSM is applicable to various tracking scenarios and achieves excellent tracking performance across multiple benchmarks, further extending the potential of SSM-like temporal motion models in multi-object tracking tasks. Code and models are publicly available at \url{https://github.com/Xavier-Lin/TrackSSM}.

Results

TaskDatasetMetricValueModel
VideoSportsMOTAssA62.4TrackSSM
VideoSportsMOTDetA88.8TrackSSM
VideoSportsMOTHOTA74.4TrackSSM
VideoSportsMOTIDF174.5TrackSSM
VideoSportsMOTMOTA96.8TrackSSM
Object TrackingSportsMOTAssA62.4TrackSSM
Object TrackingSportsMOTDetA88.8TrackSSM
Object TrackingSportsMOTHOTA74.4TrackSSM
Object TrackingSportsMOTIDF174.5TrackSSM
Object TrackingSportsMOTMOTA96.8TrackSSM
Multiple Object TrackingSportsMOTAssA62.4TrackSSM
Multiple Object TrackingSportsMOTDetA88.8TrackSSM
Multiple Object TrackingSportsMOTHOTA74.4TrackSSM
Multiple Object TrackingSportsMOTIDF174.5TrackSSM
Multiple Object TrackingSportsMOTMOTA96.8TrackSSM

Related Papers

MVA 2025 Small Multi-Object Tracking for Spotting Birds Challenge: Dataset, Methods, and Results2025-07-17YOLOv8-SMOT: An Efficient and Robust Framework for Real-Time Small Object Tracking via Slice-Assisted Training and Adaptive Association2025-07-16HiM2SAM: Enhancing SAM2 with Hierarchical Motion Estimation and Memory Optimization towards Long-term Tracking2025-07-10Robustifying 3D Perception through Least-Squares Multi-Agent Graphs Object Tracking2025-07-07UMDATrack: Unified Multi-Domain Adaptive Tracking Under Adverse Weather Conditions2025-07-01Mamba-FETrack V2: Revisiting State Space Model for Frame-Event based Visual Object Tracking2025-06-30Visual and Memory Dual Adapter for Multi-Modal Object Tracking2025-06-30R1-Track: Direct Application of MLLMs to Visual Object Tracking via Reinforcement Learning2025-06-27