TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Do Response Selection Models Really Know What's Next? Utte...

Do Response Selection Models Really Know What's Next? Utterance Manipulation Strategies for Multi-turn Response Selection

Taesun Whang, Dongyub Lee, Dongsuk Oh, Chanhee Lee, Kijong Han, Dong-hun Lee, Saebyeok Lee

2020-09-10Binary ClassificationConversational Response SelectionRetrieval
PaperPDFCode(official)

Abstract

In this paper, we study the task of selecting the optimal response given a user and system utterance history in retrieval-based multi-turn dialog systems. Recently, pre-trained language models (e.g., BERT, RoBERTa, and ELECTRA) showed significant improvements in various natural language processing tasks. This and similar response selection tasks can also be solved using such language models by formulating the tasks as dialog--response binary classification tasks. Although existing works using this approach successfully obtained state-of-the-art results, we observe that language models trained in this manner tend to make predictions based on the relatedness of history and candidates, ignoring the sequential nature of multi-turn dialog systems. This suggests that the response selection task alone is insufficient for learning temporal dependencies between utterances. To this end, we propose utterance manipulation strategies (UMS) to address this problem. Specifically, UMS consist of several strategies (i.e., insertion, deletion, and search), which aid the response selection model towards maintaining dialog coherence. Further, UMS are self-supervised methods that do not require additional annotation and thus can be easily incorporated into existing approaches. Extensive evaluation across multiple languages and models shows that UMS are highly effective in teaching dialog consistency, which leads to models pushing the state-of-the-art with significant margins on multiple public benchmark datasets.

Results

TaskDatasetMetricValueModel
Conversational Response SelectionDoubanMAP0.625UMS_BERT+
Conversational Response SelectionDoubanMRR0.664UMS_BERT+
Conversational Response SelectionDoubanP@10.499UMS_BERT+
Conversational Response SelectionDoubanR10@10.318UMS_BERT+
Conversational Response SelectionDoubanR10@20.482UMS_BERT+
Conversational Response SelectionDoubanR10@50.858UMS_BERT+
Conversational Response SelectionUbuntu Dialogue (v1, Ranking)R10@10.875UMS_BERT+
Conversational Response SelectionUbuntu Dialogue (v1, Ranking)R10@20.942UMS_BERT+
Conversational Response SelectionUbuntu Dialogue (v1, Ranking)R10@50.988UMS_BERT+
Conversational Response SelectionE-commerceR10@10.762UMS_BERT+
Conversational Response SelectionE-commerceR10@20.905UMS_BERT+
Conversational Response SelectionE-commerceR10@50.986UMS_BERT+

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17MCoT-RE: Multi-Faceted Chain-of-Thought and Re-Ranking for Training-Free Zero-Shot Composed Image Retrieval2025-07-17Developing Visual Augmented Q&A System using Scalable Vision Embedding Retrieval & Late Interaction Re-ranker2025-07-16Language-Guided Contrastive Audio-Visual Masked Autoencoder with Automatically Generated Audio-Visual-Text Triplets from Videos2025-07-16Context-Aware Search and Retrieval Over Erasure Channels2025-07-16Seq vs Seq: An Open Suite of Paired Encoders and Decoders2025-07-15