TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Two-Stage Prediction-Aware Contrastive Learning Framewor...

A Two-Stage Prediction-Aware Contrastive Learning Framework for Multi-Intent NLU

Guanhua Chen, Yutong Yao, Derek F. Wong, Lidia S. Chao

2024-05-05Semantic Frame ParsingNatural Language UnderstandingIntent DetectionData AugmentationSlot FillingPredictionContrastive Learning
PaperPDF

Abstract

Multi-intent natural language understanding (NLU) presents a formidable challenge due to the model confusion arising from multiple intents within a single utterance. While previous works train the model contrastively to increase the margin between different multi-intent labels, they are less suited to the nuances of multi-intent NLU. They ignore the rich information between the shared intents, which is beneficial to constructing a better embedding space, especially in low-data scenarios. We introduce a two-stage Prediction-Aware Contrastive Learning (PACL) framework for multi-intent NLU to harness this valuable knowledge. Our approach capitalizes on shared intent information by integrating word-level pre-training and prediction-aware contrastive fine-tuning. We construct a pre-training dataset using a word-level data augmentation strategy. Subsequently, our framework dynamically assigns roles to instances during contrastive fine-tuning while introducing a prediction-aware contrastive loss to maximize the impact of contrastive learning. We present experimental results and empirical analysis conducted on three widely used datasets, demonstrating that our method surpasses the performance of three prominent baselines on both low-data and full-data scenarios.

Results

TaskDatasetMetricValueModel
Slot FillingMixSNIPSMicro F196.8SLIM (PACL)
Slot FillingMixSNIPSMicro F196.3TFMN (PACL)
Slot FillingMixSNIPSMicro F196.2RoBERTa (PACL)
Slot FillingMixATISMicro F187.3SLIM (PACL)
Slot FillingMixATISMicro F186.7TFMN (PACL)
Slot FillingMixATISMicro F186RoBERTa (PACL)
Intent DetectionMixSNIPSAccuracy97.4TFMN (PACL)
Intent DetectionMixSNIPSAccuracy96.9SLIM (PACL)
Intent DetectionMixSNIPSAccuracy96.5RoBERTa (PACL)
Intent DetectionMixATISAccuracy82.9TFMN (PACL)
Intent DetectionMixATISAccuracy81.9SLIM (PACL)
Intent DetectionMixATISAccuracy79.1RoBERTa (PACL)

Related Papers

Multi-Strategy Improved Snake Optimizer Accelerated CNN-LSTM-Attention-Adaboost for Trajectory Prediction2025-07-21Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16Data Augmentation in Time Series Forecasting through Inverted Framework2025-07-15