TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/TrICy: Trigger-guided Data-to-text Generation with Intent ...

TrICy: Trigger-guided Data-to-text Generation with Intent aware Attention-Copy

Vibhav Agarwal, Sourav Ghosh, Harichandana BSS, Himanshu Arora, Barath Raj Kandur Raja

2024-01-25IEEE/ACM Transactions on Audio, Speech, and Language Processing 2024 1Data-to-Text GenerationText GenerationNatural Language Understanding
PaperPDF

Abstract

Data-to-text (D2T) generation is a crucial task in many natural language understanding (NLU) applications and forms the foundation of task-oriented dialog systems. In the context of conversational AI solutions that can work directly with local data on the user's device, architectures utilizing large pre-trained language models (PLMs) are impractical for on-device deployment due to a high memory footprint. To this end, we propose TrICy, a novel lightweight framework for an enhanced D2T task that generates text sequences based on the intent in context and may further be guided by user-provided triggers. We leverage an attention-copy mechanism to predict out-of-vocabulary (OOV) words accurately. Performance analyses on E2E NLG dataset (BLEU: 66.43%, ROUGE-L: 70.14%), WebNLG dataset (BLEU: Seen 64.08%, Unseen 52.35%), and our Custom dataset related to text messaging applications, showcase our architecture's effectiveness. Moreover, we show that by leveraging an optional trigger input, data-to-text generation quality increases significantly and achieves the new SOTA score of 69.29% BLEU for E2E NLG. Furthermore, our analyses show that TrICy achieves at least 24% and 3% improvement in BLEU and METEOR respectively over LLMs like GPT-3, ChatGPT, and Llama 2. We also demonstrate that in some scenarios, performance improvement due to triggers is observed even when they are absent in training.

Results

TaskDatasetMetricValueModel
Text GenerationWebNLGBLEU64.73TrICy (trK = trk* = 0.24)
Text GenerationWebNLGMETEOR45.53TrICy (trK = trk* = 0.24)
Text GenerationWebNLGNumber of parameters (M)6.2TrICy (trK = trk* = 0.24)
Text GenerationWebNLGBLEU64.08TrICy (trK = 0)
Text GenerationWebNLGMETEOR45.23TrICy (trK = 0)
Text GenerationWebNLGNumber of parameters (M)6.2TrICy (trK = 0)
Text GenerationE2E NLG ChallengeBLEU66.43TrICy (trK = 0)
Text GenerationE2E NLG ChallengeNumber of parameters (M)4.7TrICy (trK = 0)
Text GenerationE2E NLG ChallengeROUGE-L70.14TrICy (trK = 0)
Data-to-Text GenerationWebNLGBLEU64.73TrICy (trK = trk* = 0.24)
Data-to-Text GenerationWebNLGMETEOR45.53TrICy (trK = trk* = 0.24)
Data-to-Text GenerationWebNLGNumber of parameters (M)6.2TrICy (trK = trk* = 0.24)
Data-to-Text GenerationWebNLGBLEU64.08TrICy (trK = 0)
Data-to-Text GenerationWebNLGMETEOR45.23TrICy (trK = 0)
Data-to-Text GenerationWebNLGNumber of parameters (M)6.2TrICy (trK = 0)
Data-to-Text GenerationE2E NLG ChallengeBLEU66.43TrICy (trK = 0)
Data-to-Text GenerationE2E NLG ChallengeNumber of parameters (M)4.7TrICy (trK = 0)
Data-to-Text GenerationE2E NLG ChallengeROUGE-L70.14TrICy (trK = 0)

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16The Devil behind the mask: An emergent safety vulnerability of Diffusion LLMs2025-07-15Seq vs Seq: An Open Suite of Paired Encoders and Decoders2025-07-15Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15Vision Language Action Models in Robotic Manipulation: A Systematic Review2025-07-14Exploiting Leaderboards for Large-Scale Distribution of Malicious Models2025-07-11CLI-RAG: A Retrieval-Augmented Framework for Clinically Structured and Context Aware Text Generation with LLMs2025-07-09