TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/It Takes Two: Accurate Gait Recognition in the Wild via Cr...

It Takes Two: Accurate Gait Recognition in the Wild via Cross-granularity Alignment

Jinkai Zheng, Xinchen Liu, Boyue Zhang, Chenggang Yan, Jiyong Zhang, Wu Liu, Yongdong Zhang

2024-11-16Gait Recognition in the WildGait RecognitionHuman Parsing
PaperPDFCode

Abstract

Existing studies for gait recognition primarily utilized sequences of either binary silhouette or human parsing to encode the shapes and dynamics of persons during walking. Silhouettes exhibit accurate segmentation quality and robustness to environmental variations, but their low information entropy may result in sub-optimal performance. In contrast, human parsing provides fine-grained part segmentation with higher information entropy, but the segmentation quality may deteriorate due to the complex environments. To discover the advantages of silhouette and parsing and overcome their limitations, this paper proposes a novel cross-granularity alignment gait recognition method, named XGait, to unleash the power of gait representations of different granularity. To achieve this goal, the XGait first contains two branches of backbone encoders to map the silhouette sequences and the parsing sequences into two latent spaces, respectively. Moreover, to explore the complementary knowledge across the features of two representations, we design the Global Cross-granularity Module (GCM) and the Part Cross-granularity Module (PCM) after the two encoders. In particular, the GCM aims to enhance the quality of parsing features by leveraging global features from silhouettes, while the PCM aligns the dynamics of human parts between silhouette and parsing features using the high information entropy in parsing sequences. In addition, to effectively guide the alignment of two representations with different granularity at the part level, an elaborate-designed learnable division mechanism is proposed for the parsing features. Comprehensive experiments on two large-scale gait datasets not only show the superior performance of XGait with the Rank-1 accuracy of 80.5% on Gait3D and 88.3% CCPG but also reflect the robustness of the learned features even under challenging conditions like occlusions and cloth changes.

Related Papers

Mind the Gap: Bridging Occlusion in Gait Recognition via Residual Gap Correction2025-07-15On Denoising Walking Videos for Gait Recognition2025-05-24ExoGait-MS: Learning Periodic Dynamics with Multi-Scale Graph Network for Exoskeleton Gait Recognition2025-05-23BiggerGait: Unlocking Gait Recognition with Layer-wise Representations from Large Vision Models2025-05-23Exploring Generalized Gait Recognition: Reducing Redundancy and Noise within Indoor and Outdoor Datasets2025-05-21OptiGait-LGBM: An Efficient Approach of Gait-based Person Re-identification in Non-Overlapping Regions2025-05-10Database-Agnostic Gait Enrollment using SetTransformers2025-05-05CVVNet: A Cross-Vertical-View Network for Gait Recognition2025-05-03