TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Image Referenced Sketch Colorization Based on Animation Cr...

Image Referenced Sketch Colorization Based on Animation Creation Workflow

Dingkun Yan, Xinrui Wang, Zhuoru Li, Suguru Saito, Yusuke Iwasawa, Yutaka Matsuo, Jiaxian Guo

2025-02-27CVPR 2025 1Colorization
PaperPDFCodeCode(official)

Abstract

Sketch colorization plays an important role in animation and digital illustration production tasks. However, existing methods still meet problems in that text-guided methods fail to provide accurate color and style reference, hint-guided methods still involve manual operation, and image-referenced methods are prone to cause artifacts. To address these limitations, we propose a diffusion-based framework inspired by real-world animation production workflows. Our approach leverages the sketch as the spatial guidance and an RGB image as the color reference, and separately extracts foreground and background from the reference image with spatial masks. Particularly, we introduce a split cross-attention mechanism with LoRA (Low-Rank Adaptation) modules. They are trained separately with foreground and background regions to control the corresponding embeddings for keys and values in cross-attention. This design allows the diffusion model to integrate information from foreground and background independently, preventing interference and eliminating the spatial artifacts. During inference, we design switchable inference modes for diverse use scenarios by changing modules activated in the framework. Extensive qualitative and quantitative experiments, along with user studies, demonstrate our advantages over existing methods in generating high-qualigy artifact-free results with geometric mismatched references. Ablation studies further confirm the effectiveness of each component. Codes are available at https://github.com/ tellurion-kanata/colorizeDiffusion.

Related Papers

MTSIC: Multi-stage Transformer-based GAN for Spectral Infrared Image Colorization2025-06-21Self-supervised Feature Extraction for Enhanced Ball Detection on Soccer Robots2025-06-20Exploiting the Exact Denoising Posterior Score in Training-Free Guidance of Diffusion Models2025-06-16SSIMBaD: Sigma Scaling with SSIM-Guided Balanced Diffusion for AnimeFace Colorization2025-06-04Restoring Real-World Images with an Internal Detail Enhancement Diffusion Model2025-05-24Leveraging the Powerful Attention of a Pre-trained Diffusion Model for Exemplar-based Image Colorization2025-05-21Controllable Image Colorization with Instance-aware Texts and Masks2025-05-13ColorVein: Colorful Cancelable Vein Biometrics2025-04-19