Ming Ding, Wendi Zheng, Wenyi Hong, Jie Tang
The development of the transformer-based text-to-image models are impeded by its slow generation and complexity for high-resolution images. In this work, we put forward a solution based on hierarchical transformers and local parallel auto-regressive generation. We pretrain a 6B-parameter transformer with a simple and flexible self-supervised task, Cross-modal general language model (CogLM), and finetune it for fast super-resolution. The new text-to-image system, CogView2, shows very competitive generation compared to concurrent state-of-the-art DALL-E-2, and naturally supports interactive text-guided editing on images.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Image Generation | COCO (Common Objects in Context) | FID | 17.7 | CogView2(6B, Finetuned) |
| Image Generation | COCO (Common Objects in Context) | FID | 24 | CogView2(6B, Finetuned) |
| Text-to-Image Generation | COCO (Common Objects in Context) | FID | 17.7 | CogView2(6B, Finetuned) |
| Text-to-Image Generation | COCO (Common Objects in Context) | FID | 24 | CogView2(6B, Finetuned) |
| 10-shot image generation | COCO (Common Objects in Context) | FID | 17.7 | CogView2(6B, Finetuned) |
| 10-shot image generation | COCO (Common Objects in Context) | FID | 24 | CogView2(6B, Finetuned) |
| 1 Image, 2*2 Stitchi | COCO (Common Objects in Context) | FID | 17.7 | CogView2(6B, Finetuned) |
| 1 Image, 2*2 Stitchi | COCO (Common Objects in Context) | FID | 24 | CogView2(6B, Finetuned) |