TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning Hierarchical Prompt with Structured Linguistic Kn...

Learning Hierarchical Prompt with Structured Linguistic Knowledge for Vision-Language Models

Yubin Wang, Xinyang Jiang, De Cheng, Dongsheng Li, Cairong Zhao

2023-12-11Prompt Engineering
PaperPDFCode(official)Code

Abstract

Prompt learning has become a prevalent strategy for adapting vision-language foundation models to downstream tasks. As large language models (LLMs) have emerged, recent studies have explored the use of category-related descriptions as input to enhance prompt effectiveness. Nevertheless, conventional descriptions fall short of structured information that effectively represents the interconnections among entities or attributes linked to a particular category. To address this limitation and prioritize harnessing structured knowledge, this paper advocates for leveraging LLMs to build a graph for each description to model the entities and attributes describing the category, as well as their correlations. Preexisting prompt tuning methods exhibit inadequacies in managing this structured knowledge. Consequently, we propose a novel approach called Hierarchical Prompt Tuning (HPT), which enables simultaneous modeling of both structured and conventional linguistic knowledge. Specifically, we introduce a relationship-guided attention module to capture pair-wise associations among entities and attributes for low-level prompt learning. In addition, by incorporating high-level and global-level prompts modeling overall semantics, the proposed hierarchical structure forges cross-level interlinks and empowers the model to handle more complex and long-term relationships. Extensive experiments demonstrate that our HPT shows strong effectiveness and generalizes much better than existing SOTA methods. Our code is available at https://github.com/Vill-Lab/2024-AAAI-HPT.

Results

TaskDatasetMetricValueModel
Prompt EngineeringImageNet-RTop-1 accuracy %77.38HPT
Prompt EngineeringStanford CarsHarmonic mean75.57HPT
Prompt EngineeringOxford 102 FlowerHarmonic mean87.16HPT
Prompt EngineeringEuroSATHarmonic mean84.82HPT
Prompt EngineeringOxford-IIIT Pet DatasetHarmonic mean96.71HPT
Prompt EngineeringImageNet-STop-1 accuracy %49.36HPT
Prompt EngineeringDTDHarmonic mean72.16HPT
Prompt EngineeringUCF101Harmonic mean83.16HPT
Prompt EngineeringFood-101Harmonic mean91.01HPT
Prompt EngineeringCaltech-101Harmonic mean96.65HPT
Prompt EngineeringImageNetHarmonic mean74.17HPT
Prompt EngineeringFGVC-AircraftHarmonic mean40.28HPT
Prompt EngineeringSUN397Harmonic mean80.88HPT
Prompt EngineeringImageNet-ATop-1 accuracy %50.85HPT
Prompt EngineeringImageNet V2Top-1 accuracy %65.25HPT

Related Papers

Leveraging Language Prior for Infrared Small Target Detection2025-07-17Emotional Support with LLM-based Empathetic Dialogue Generation2025-07-17Prompt Engineering in Segment Anything Model: Methodologies, Applications, and Emerging Challenges2025-07-13AdaptaGen: Domain-Specific Image Generation through Hierarchical Semantic Optimization Framework2025-07-08Helping CLIP See Both the Forest and the Trees: A Decomposition and Description Approach2025-07-04State and Memory is All You Need for Robust and Reliable AI Agents2025-06-30Prompt Mechanisms in Medical Imaging: A Comprehensive Survey2025-06-28Fine-Tuning and Prompt Engineering of LLMs, for the Creation of Multi-Agent AI for Addressing Sustainable Protein Production Challenges2025-06-25