TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Drawing the Same Bounding Box Twice? Coping Noisy Annotati...

Drawing the Same Bounding Box Twice? Coping Noisy Annotations in Object Detection with Repeated Labels

David Tschirschwitz, Christian Benz, Morris Florek, Henrik Norderhus, Benno Stein, Volker Rodehorst

2023-09-18Instance Segmentationobject-detectionObject Detection
PaperPDFCode(official)

Abstract

The reliability of supervised machine learning systems depends on the accuracy and availability of ground truth labels. However, the process of human annotation, being prone to error, introduces the potential for noisy labels, which can impede the practicality of these systems. While training with noisy labels is a significant consideration, the reliability of test data is also crucial to ascertain the dependability of the results. A common approach to addressing this issue is repeated labeling, where multiple annotators label the same example, and their labels are combined to provide a better estimate of the true label. In this paper, we propose a novel localization algorithm that adapts well-established ground truth estimation methods for object detection and instance segmentation tasks. The key innovation of our method lies in its ability to transform combined localization and classification tasks into classification-only problems, thus enabling the application of techniques such as Expectation-Maximization (EM) or Majority Voting (MJV). Although our main focus is the aggregation of unique ground truth for test data, our algorithm also shows superior performance during training on the TexBiG dataset, surpassing both noisy label training and label aggregation using Weighted Boxes Fusion (WBF). Our experiments indicate that the benefits of repeated labels emerge under specific dataset and annotation configurations. The key factors appear to be (1) dataset complexity, the (2) annotator consistency, and (3) the given annotation budget constraints.

Results

TaskDatasetMetricValueModel
Object DetectionTexBiG 2023 testmAP@0.5:0.95:0.0549.89DetectoRS + LAEM
3DTexBiG 2023 testmAP@0.5:0.95:0.0549.89DetectoRS + LAEM
Instance SegmentationTexBiG 2023 testmAP@0.5:0.95:0.0544.06DetectoRS + LAEM
2D ClassificationTexBiG 2023 testmAP@0.5:0.95:0.0549.89DetectoRS + LAEM
2D Object DetectionTexBiG 2023 testmAP@0.5:0.95:0.0549.89DetectoRS + LAEM
16kTexBiG 2023 testmAP@0.5:0.95:0.0549.89DetectoRS + LAEM

Related Papers

SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17RS-TinyNet: Stage-wise Feature Fusion Network for Detecting Tiny Objects in Remote Sensing Images2025-07-17Decoupled PROB: Decoupled Query Initialization Tasks and Objectness-Class Learning for Open World Object Detection2025-07-17Dual LiDAR-Based Traffic Movement Count Estimation at a Signalized Intersection: Deployment, Data Collection, and Preliminary Analysis2025-07-17Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Tomato Multi-Angle Multi-Pose Dataset for Fine-Grained Phenotyping2025-07-15Beyond Appearance: Geometric Cues for Robust Video Instance Segmentation2025-07-08