TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MISCA: A Joint Model for Multiple Intent Detection and Slo...

MISCA: A Joint Model for Multiple Intent Detection and Slot Filling with Intent-Slot Co-Attention

Thinh Pham, Chi Tran, Dat Quoc Nguyen

2023-12-10Semantic Frame ParsingIntent Detectionslot-fillingSlot Fillinggraph construction
PaperPDFCode(official)

Abstract

The research study of detecting multiple intents and filling slots is becoming more popular because of its relevance to complicated real-world situations. Recent advanced approaches, which are joint models based on graphs, might still face two potential issues: (i) the uncertainty introduced by constructing graphs based on preliminary intents and slots, which may transfer intent-slot correlation information to incorrect label node destinations, and (ii) direct incorporation of multiple intent labels for each token w.r.t. token-level intent voting might potentially lead to incorrect slot predictions, thereby hurting the overall performance. To address these two issues, we propose a joint model named MISCA. Our MISCA introduces an intent-slot co-attention mechanism and an underlying layer of label attention mechanism. These mechanisms enable MISCA to effectively capture correlations between intents and slot labels, eliminating the need for graph construction. They also facilitate the transfer of correlation information in both directions: from intents to slots and from slots to intents, through multiple levels of label-specific representations, without relying on token-level intent information. Experimental results show that MISCA outperforms previous models, achieving new state-of-the-art overall accuracy performances on two benchmark datasets MixATIS and MixSNIPS. This highlights the effectiveness of our attention mechanisms.

Results

TaskDatasetMetricValueModel
Slot FillingMixSNIPSMicro F195.2MISCA
Slot FillingMixATISMicro F190.5MISCA
Intent DetectionMixSNIPSAccuracy97.3MISCA
Intent DetectionMixATISAccuracy76.7MISCA

Related Papers

Efficiently Constructing Sparse Navigable Graphs2025-07-17NGTM: Substructure-based Neural Graph Topic Model for Interpretable Graph Generation2025-07-17Flippi: End To End GenAI Assistant for E-Commerce2025-07-08Irec: A Metacognitive Scaffolding for Self-Regulated Learning through Just-in-Time Insight Recall: A Conceptual Framework and System Prototype2025-06-25Call Me Maybe: Enhancing JavaScript Call Graph Construction using Graph Neural Networks2025-06-22CORE-KG: An LLM-Driven Knowledge Graph Construction Framework for Human Smuggling Networks2025-06-20An Interdisciplinary Review of Commonsense Reasoning and Intent Detection2025-06-16SemanticST: Spatially Informed Semantic Graph Learning for Clustering, Integration, and Scalable Analysis of Spatial Transcriptomics2025-06-13