Hai-Ming Xu, Lingqiao Liu, Qiuchen Bian, Zhen Yang
Semi-supervised semantic segmentation requires the model to effectively propagate the label information from limited annotated images to unlabeled ones. A challenge for such a per-pixel prediction task is the large intra-class variation, i.e., regions belonging to the same class may exhibit a very different appearance even in the same picture. This diversity will make the label propagation hard from pixels to pixels. To address this problem, we propose a novel approach to regularize the distribution of within-class features to ease label propagation difficulty. Specifically, our approach encourages the consistency between the prediction from a linear predictor and the output from a prototype-based predictor, which implicitly encourages features from the same pseudo-class to be close to at least one within-class prototype while staying far from the other between-class prototypes. By further incorporating CutMix operations and a carefully-designed prototype maintenance strategy, we create a semi-supervised semantic segmentation algorithm that demonstrates superior performance over the state-of-the-art methods from extensive experimental evaluation on both Pascal VOC and Cityscapes benchmarks.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Semantic Segmentation | Pascal VOC 2012 6.25% labeled | Validation mIoU | 78.6 | PCR (DeepLab v3+ with ResNet-101 pretraind on ImageNet-1K) |
| Semantic Segmentation | PASCAL VOC 2012 25% labeled | Validation mIoU | 80.78 | PCR (DeepLab v3+ with ResNet-101 pretraind on ImageNet-1K) |
| 10-shot image generation | Pascal VOC 2012 6.25% labeled | Validation mIoU | 78.6 | PCR (DeepLab v3+ with ResNet-101 pretraind on ImageNet-1K) |
| 10-shot image generation | PASCAL VOC 2012 25% labeled | Validation mIoU | 80.78 | PCR (DeepLab v3+ with ResNet-101 pretraind on ImageNet-1K) |