TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Identity-Preserving Text-to-Video Generation by Frequency ...

Identity-Preserving Text-to-Video Generation by Frequency Decomposition

Shenghai Yuan, Jinfa Huang, Xianyi He, Yunyuan Ge, Yujun Shi, Liuhan Chen, Jiebo Luo, Li Yuan

2024-11-26CVPR 2025 1Text-to-Video GenerationOpen-Domain Subject-to-VideoHuman-Domain Subject-to-VideoImage to Video GenerationVideo Generation
PaperPDFCode(official)

Abstract

Identity-preserving text-to-video (IPT2V) generation aims to create high-fidelity videos with consistent human identity. It is an important task in video generation but remains an open problem for generative models. This paper pushes the technical frontier of IPT2V in two directions that have not been resolved in literature: (1) A tuning-free pipeline without tedious case-by-case finetuning, and (2) A frequency-aware heuristic identity-preserving DiT-based control scheme. We propose ConsisID, a tuning-free DiT-based controllable IPT2V model to keep human identity consistent in the generated video. Inspired by prior findings in frequency analysis of diffusion transformers, it employs identity-control signals in the frequency domain, where facial features can be decomposed into low-frequency global features and high-frequency intrinsic features. First, from a low-frequency perspective, we introduce a global facial extractor, which encodes reference images and facial key points into a latent space, generating features enriched with low-frequency information. These features are then integrated into shallow layers of the network to alleviate training challenges associated with DiT. Second, from a high-frequency perspective, we design a local facial extractor to capture high-frequency details and inject them into transformer blocks, enhancing the model's ability to preserve fine-grained features. We propose a hierarchical training strategy to leverage frequency information for identity preservation, transforming a vanilla pre-trained video generation model into an IPT2V model. Extensive experiments demonstrate that our frequency-aware heuristic scheme provides an optimal control solution for DiT-based models. Thanks to this scheme, our ConsisID generates high-quality, identity-preserving videos, making strides towards more effective IPT2V. Code: https://github.com/PKU-YuanGroup/ConsisID.

Results

TaskDatasetMetricValueModel
VideoOpenS2V-EvalAesthetics0.446Kling 1.6
VideoOpenS2V-EvalFaceSim0.401Kling 1.6
VideoOpenS2V-EvalGmeScore0.662Kling 1.6
VideoOpenS2V-EvalMotion0.416Kling 1.6
VideoOpenS2V-EvalNaturalScore0.7906Kling 1.6
VideoOpenS2V-EvalNexusScore0.4592Kling 1.6
VideoOpenS2V-EvalTotal Score0.5446Kling 1.6
VideoOpenS2V-EvalAesthetics0.4687Pika 2.1
VideoOpenS2V-EvalFaceSim0.308Pika 2.1
VideoOpenS2V-EvalGmeScore0.6921Pika 2.1
VideoOpenS2V-EvalMotion0.247Pika 2.1
VideoOpenS2V-EvalNaturalScore0.6979Pika 2.1
VideoOpenS2V-EvalNexusScore0.4541Pika 2.1
VideoOpenS2V-EvalTotal Score0.4888Pika 2.1
VideoOpenS2V-EvalAesthetics0.4147Vidu 2.0
VideoOpenS2V-EvalFaceSim0.3511Vidu 2.0
VideoOpenS2V-EvalGmeScore0.6757Vidu 2.0
VideoOpenS2V-EvalMotion0.1352Vidu 2.0
VideoOpenS2V-EvalNaturalScore0.7144Vidu 2.0
VideoOpenS2V-EvalNexusScore0.4355Vidu 2.0
VideoOpenS2V-EvalTotal Score0.4759Vidu 2.0
Video GenerationOpenS2V-EvalAesthetics0.446Kling 1.6
Video GenerationOpenS2V-EvalFaceSim0.401Kling 1.6
Video GenerationOpenS2V-EvalGmeScore0.662Kling 1.6
Video GenerationOpenS2V-EvalMotion0.416Kling 1.6
Video GenerationOpenS2V-EvalNaturalScore0.7906Kling 1.6
Video GenerationOpenS2V-EvalNexusScore0.4592Kling 1.6
Video GenerationOpenS2V-EvalTotal Score0.5446Kling 1.6
Video GenerationOpenS2V-EvalAesthetics0.4687Pika 2.1
Video GenerationOpenS2V-EvalFaceSim0.308Pika 2.1
Video GenerationOpenS2V-EvalGmeScore0.6921Pika 2.1
Video GenerationOpenS2V-EvalMotion0.247Pika 2.1
Video GenerationOpenS2V-EvalNaturalScore0.6979Pika 2.1
Video GenerationOpenS2V-EvalNexusScore0.4541Pika 2.1
Video GenerationOpenS2V-EvalTotal Score0.4888Pika 2.1
Video GenerationOpenS2V-EvalAesthetics0.4147Vidu 2.0
Video GenerationOpenS2V-EvalFaceSim0.3511Vidu 2.0
Video GenerationOpenS2V-EvalGmeScore0.6757Vidu 2.0
Video GenerationOpenS2V-EvalMotion0.1352Vidu 2.0
Video GenerationOpenS2V-EvalNaturalScore0.7144Vidu 2.0
Video GenerationOpenS2V-EvalNexusScore0.4355Vidu 2.0
Video GenerationOpenS2V-EvalTotal Score0.4759Vidu 2.0
1 Image, 2*2 StitchiOpenS2V-EvalAesthetics0.446Kling 1.6
1 Image, 2*2 StitchiOpenS2V-EvalFaceSim0.401Kling 1.6
1 Image, 2*2 StitchiOpenS2V-EvalGmeScore0.662Kling 1.6
1 Image, 2*2 StitchiOpenS2V-EvalMotion0.416Kling 1.6
1 Image, 2*2 StitchiOpenS2V-EvalNaturalScore0.7906Kling 1.6
1 Image, 2*2 StitchiOpenS2V-EvalNexusScore0.4592Kling 1.6
1 Image, 2*2 StitchiOpenS2V-EvalTotal Score0.5446Kling 1.6
1 Image, 2*2 StitchiOpenS2V-EvalAesthetics0.4687Pika 2.1
1 Image, 2*2 StitchiOpenS2V-EvalFaceSim0.308Pika 2.1
1 Image, 2*2 StitchiOpenS2V-EvalGmeScore0.6921Pika 2.1
1 Image, 2*2 StitchiOpenS2V-EvalMotion0.247Pika 2.1
1 Image, 2*2 StitchiOpenS2V-EvalNaturalScore0.6979Pika 2.1
1 Image, 2*2 StitchiOpenS2V-EvalNexusScore0.4541Pika 2.1
1 Image, 2*2 StitchiOpenS2V-EvalTotal Score0.4888Pika 2.1
1 Image, 2*2 StitchiOpenS2V-EvalAesthetics0.4147Vidu 2.0
1 Image, 2*2 StitchiOpenS2V-EvalFaceSim0.3511Vidu 2.0
1 Image, 2*2 StitchiOpenS2V-EvalGmeScore0.6757Vidu 2.0
1 Image, 2*2 StitchiOpenS2V-EvalMotion0.1352Vidu 2.0
1 Image, 2*2 StitchiOpenS2V-EvalNaturalScore0.7144Vidu 2.0
1 Image, 2*2 StitchiOpenS2V-EvalNexusScore0.4355Vidu 2.0
1 Image, 2*2 StitchiOpenS2V-EvalTotal Score0.4759Vidu 2.0
Image to Video GenerationOpenS2V-EvalAesthetics0.446Kling 1.6
Image to Video GenerationOpenS2V-EvalFaceSim0.401Kling 1.6
Image to Video GenerationOpenS2V-EvalGmeScore0.662Kling 1.6
Image to Video GenerationOpenS2V-EvalMotion0.416Kling 1.6
Image to Video GenerationOpenS2V-EvalNaturalScore0.7906Kling 1.6
Image to Video GenerationOpenS2V-EvalNexusScore0.4592Kling 1.6
Image to Video GenerationOpenS2V-EvalTotal Score0.5446Kling 1.6
Image to Video GenerationOpenS2V-EvalAesthetics0.4687Pika 2.1
Image to Video GenerationOpenS2V-EvalFaceSim0.308Pika 2.1
Image to Video GenerationOpenS2V-EvalGmeScore0.6921Pika 2.1
Image to Video GenerationOpenS2V-EvalMotion0.247Pika 2.1
Image to Video GenerationOpenS2V-EvalNaturalScore0.6979Pika 2.1
Image to Video GenerationOpenS2V-EvalNexusScore0.4541Pika 2.1
Image to Video GenerationOpenS2V-EvalTotal Score0.4888Pika 2.1
Image to Video GenerationOpenS2V-EvalAesthetics0.4147Vidu 2.0
Image to Video GenerationOpenS2V-EvalFaceSim0.3511Vidu 2.0
Image to Video GenerationOpenS2V-EvalGmeScore0.6757Vidu 2.0
Image to Video GenerationOpenS2V-EvalMotion0.1352Vidu 2.0
Image to Video GenerationOpenS2V-EvalNaturalScore0.7144Vidu 2.0
Image to Video GenerationOpenS2V-EvalNexusScore0.4355Vidu 2.0
Image to Video GenerationOpenS2V-EvalTotal Score0.4759Vidu 2.0

Related Papers

LoViC: Efficient Long Video Generation with Context Compression2025-07-17World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17Leveraging Pre-Trained Visual Models for AI-Generated Video Detection2025-07-17Taming Diffusion Transformer for Real-Time Mobile Video Generation2025-07-17$I^{2}$-World: Intra-Inter Tokenization for Efficient Dynamic 4D Scene Forecasting2025-07-12Lumos-1: On Autoregressive Video Generation from a Unified Model Perspective2025-07-11Scaling RL to Long Videos2025-07-10Martian World Models: Controllable Video Synthesis with Physically Accurate 3D Reconstructions2025-07-10