TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning to Memorize Entailment and Discourse Relations fo...

Learning to Memorize Entailment and Discourse Relations for Persona-Consistent Dialogues

Ruijun Chen, Jin Wang, Liang-Chih Yu, Xuejie Zhang

2023-01-12Dialogue Generation
PaperPDFCode(official)

Abstract

Maintaining engagement and consistency is particularly important in dialogue systems. Existing works have improved the performance of dialogue systems by intentionally learning interlocutor personas with sophisticated network structures. One issue with this approach is that it requires more personal corpora with annotations. Additionally, these models typically perform the next utterance prediction to generate a response but neglect the discourse coherence in the entire conversation. To address these issues, this study proposes a method of learning to memorize entailment and discourse relations for persona-consistent dialogue tasks. Entailment text pairs in natural language inference dataset were applied to learn latent entailment relations as external memories by premise-to-hypothesis generation task. Furthermore, an internal memory with a similar architecture was applied to the discourse information in the dialogue. Placing orthogonality restrictions on these two memory spaces ensures that the latent entailment relations remain dialogue-independent. Both memories collaborate to obtain entailment and discourse representation for the generation, allowing a deeper understanding of both consistency and coherence. Experiments on two large public datasets, PersonaChat and DSTC7-AVSD, demonstrated the effectiveness of the proposed method. Both automatic and human evaluations indicate that the proposed model outperforms several strong baselines in terms of both persona consistency and response coherence. Our source code is available at https://github.com/Chenrj233/LMEDR.

Results

TaskDatasetMetricValueModel
DialoguePersona-ChatAvg F121.99LMEDR
Text GenerationPersona-ChatAvg F121.99LMEDR
ChatbotPersona-ChatAvg F121.99LMEDR
Dialogue GenerationPersona-ChatAvg F121.99LMEDR

Related Papers

Emotional Support with LLM-based Empathetic Dialogue Generation2025-07-17ZipVoice-Dialog: Non-Autoregressive Spoken Dialogue Generation with Flow Matching2025-07-12SDialog: A Python Toolkit for Synthetic Dialogue Generation and Analysis2025-06-12Enhancing Medical Dialogue Generation through Knowledge Refinement and Dynamic Prompt Adjustment2025-06-12Proactive Assistant Dialogue Generation from Streaming Egocentric Videos2025-06-06ConsistentChat: Building Skeleton-Guided Consistent Dialogues for Large Language Models from Scratch2025-06-04CoVoMix2: Advancing Zero-Shot Dialogue Generation with Fully Non-Autoregressive Flow Matching2025-06-01Adaptive-VP: A Framework for LLM-Based Virtual Patients that Adapts to Trainees' Dialogue to Facilitate Nurse Communication Training2025-05-31