TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Biasing Like Human: A Cognitive Bias Framework for Scene G...

Biasing Like Human: A Cognitive Bias Framework for Scene Graph Generation

Xiaoguang Chang, Teng Wang, Changyin Sun, Wenzhe Cai

2022-03-17Scene Graph GenerationPredicate ClassificationGraph Generation
PaperPDFCode(official)

Abstract

Scene graph generation is a sophisticated task because there is no specific recognition pattern (e.g., "looking at" and "near" have no conspicuous difference concerning vision, whereas "near" could occur between entities with different morphology). Thus some scene graph generation methods are trapped into most frequent relation predictions caused by capricious visual features and trivial dataset annotations. Therefore, recent works emphasized the "unbiased" approaches to balance predictions for a more informative scene graph. However, human's quick and accurate judgments over relations between numerous objects should be attributed to "bias" (i.e., experience and linguistic knowledge) rather than pure vision. To enhance the model capability, inspired by the "cognitive bias" mechanism, we propose a novel 3-paradigms framework that simulates how humans incorporate the label linguistic features as guidance of vision-based representations to better mine hidden relation patterns and alleviate noisy visual propagation. Our framework is model-agnostic to any scene graph model. Comprehensive experiments prove our framework outperforms baseline modules in several metrics with minimum parameters increment and achieves new SOTA performance on Visual Genome dataset.

Results

TaskDatasetMetricValueModel
Scene ParsingVisual Genomemean Recall @10017.24C-bias
Scene ParsingVisual Genomemean Recall @2011.63C-bias
2D Semantic SegmentationVisual Genomemean Recall @10017.24C-bias
2D Semantic SegmentationVisual Genomemean Recall @2011.63C-bias
Scene Graph GenerationVisual Genomemean Recall @10017.24C-bias
Scene Graph GenerationVisual Genomemean Recall @2011.63C-bias

Related Papers

NGTM: Substructure-based Neural Graph Topic Model for Interpretable Graph Generation2025-07-17GNN-CNN: An Efficient Hybrid Model of Convolutional and Graph Neural Networks for Text Representation2025-07-10SPADE: Spatial-Aware Denoising Network for Open-vocabulary Panoptic Scene Graph Generation with Long- and Local-range Context Reasoning2025-07-08GDGB: A Benchmark for Generative Dynamic Text-Attributed Graph Learning2025-07-04CoPa-SG: Dense Scene Graphs with Parametric and Proto-Relations2025-06-26CAT-SG: A Large Dynamic Scene Graph Dataset for Fine-Grained Understanding of Cataract Surgery2025-06-26HOIverse: A Synthetic Scene Graph Dataset With Human Object Interactions2025-06-24DiscoSG: Towards Discourse-Level Text Scene Graph Parsing through Iterative Graph Refinement2025-06-18