TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Semi-supervised Human Pose Estimation in Art-historical Im...

Semi-supervised Human Pose Estimation in Art-historical Images

Matthias Springstein, Stefanie Schneider, Christian Althaus, Ralph Ewerth

2022-07-06Style Transfer2D Human Pose EstimationPose EstimationKeypoint DetectionRetrievalSemi-Supervised Human Pose Estimation
PaperPDFCode(official)

Abstract

Gesture as language of non-verbal communication has been theoretically established since the 17th century. However, its relevance for the visual arts has been expressed only sporadically. This may be primarily due to the sheer overwhelming amount of data that traditionally had to be processed by hand. With the steady progress of digitization, though, a growing number of historical artifacts have been indexed and made available to the public, creating a need for automatic retrieval of art-historical motifs with similar body constellations or poses. Since the domain of art differs significantly from existing real-world data sets for human pose estimation due to its style variance, this presents new challenges. In this paper, we propose a novel approach to estimate human poses in art-historical images. In contrast to previous work that attempts to bridge the domain gap with pre-trained models or through style transfer, we suggest semi-supervised learning for both object and keypoint detection. Furthermore, we introduce a novel domain-specific art data set that includes both bounding box and keypoint annotations of human figures. Our approach achieves significantly better results than methods that use pre-trained models or style transfer.

Results

TaskDatasetMetricValueModel
Pose EstimationPoPArtmAP52.58HRNet-W32
Pose EstimationPoPArtmAP@0.563.92HRNet-W32
Pose EstimationPoPArtmAP@0.7557.35HRNet-W32
Pose EstimationPoPArtmAP29.71HRNet-W32 (trained on PeopleArt)
Pose EstimationPoPArtmAP@0.536.37HRNet-W32 (trained on PeopleArt)
Pose EstimationPoPArtmAP@0.7532.72HRNet-W32 (trained on PeopleArt)
Pose EstimationPoPArtmAP25.25HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
Pose EstimationPoPArtmAP@0.531.73HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
Pose EstimationPoPArtmAP@0.7528.1HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
Pose EstimationPoPArtmAP25.18HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
Pose EstimationPoPArtmAP@0.531.67HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
Pose EstimationPoPArtmAP@0.7528.13HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
Pose EstimationPoPArtmAP24.13HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
Pose EstimationPoPArtmAP@0.530.52HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
Pose EstimationPoPArtmAP@0.7526.65HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
3DPoPArtmAP52.58HRNet-W32
3DPoPArtmAP@0.563.92HRNet-W32
3DPoPArtmAP@0.7557.35HRNet-W32
3DPoPArtmAP29.71HRNet-W32 (trained on PeopleArt)
3DPoPArtmAP@0.536.37HRNet-W32 (trained on PeopleArt)
3DPoPArtmAP@0.7532.72HRNet-W32 (trained on PeopleArt)
3DPoPArtmAP25.25HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
3DPoPArtmAP@0.531.73HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
3DPoPArtmAP@0.7528.1HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
3DPoPArtmAP25.18HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
3DPoPArtmAP@0.531.67HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
3DPoPArtmAP@0.7528.13HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
3DPoPArtmAP24.13HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
3DPoPArtmAP@0.530.52HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
3DPoPArtmAP@0.7526.65HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP52.58HRNet-W32
2D Human Pose EstimationPoPArtmAP@0.563.92HRNet-W32
2D Human Pose EstimationPoPArtmAP@0.7557.35HRNet-W32
2D Human Pose EstimationPoPArtmAP29.71HRNet-W32 (trained on PeopleArt)
2D Human Pose EstimationPoPArtmAP@0.536.37HRNet-W32 (trained on PeopleArt)
2D Human Pose EstimationPoPArtmAP@0.7532.72HRNet-W32 (trained on PeopleArt)
2D Human Pose EstimationPoPArtmAP25.25HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP@0.531.73HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP@0.7528.1HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP25.18HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP@0.531.67HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP@0.7528.13HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP24.13HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP@0.530.52HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
2D Human Pose EstimationPoPArtmAP@0.7526.65HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP52.58HRNet-W32
Multi-Person Pose EstimationPoPArtmAP@0.563.92HRNet-W32
Multi-Person Pose EstimationPoPArtmAP@0.7557.35HRNet-W32
Multi-Person Pose EstimationPoPArtmAP29.71HRNet-W32 (trained on PeopleArt)
Multi-Person Pose EstimationPoPArtmAP@0.536.37HRNet-W32 (trained on PeopleArt)
Multi-Person Pose EstimationPoPArtmAP@0.7532.72HRNet-W32 (trained on PeopleArt)
Multi-Person Pose EstimationPoPArtmAP25.25HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP@0.531.73HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP@0.7528.1HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP25.18HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP@0.531.67HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP@0.7528.13HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP24.13HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP@0.530.52HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
Multi-Person Pose EstimationPoPArtmAP@0.7526.65HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP52.58HRNet-W32
1 Image, 2*2 StitchiPoPArtmAP@0.563.92HRNet-W32
1 Image, 2*2 StitchiPoPArtmAP@0.7557.35HRNet-W32
1 Image, 2*2 StitchiPoPArtmAP29.71HRNet-W32 (trained on PeopleArt)
1 Image, 2*2 StitchiPoPArtmAP@0.536.37HRNet-W32 (trained on PeopleArt)
1 Image, 2*2 StitchiPoPArtmAP@0.7532.72HRNet-W32 (trained on PeopleArt)
1 Image, 2*2 StitchiPoPArtmAP25.25HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP@0.531.73HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP@0.7528.1HRNet-W32 (trained on COCO 2017 with 0 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP25.18HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP@0.531.67HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP@0.7528.13HRNet-W32 (trained on COCO 2017 with 100 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP24.13HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP@0.530.52HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)
1 Image, 2*2 StitchiPoPArtmAP@0.7526.65HRNet-W32 (trained on COCO 2017 with 50 % style-transferred Material)

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17