TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Hierarchical Pre-training for Sequence Labelling in Spoken...

Hierarchical Pre-training for Sequence Labelling in Spoken Dialog

Emile Chapuis, Pierre Colombo, Matteo Manica, Matthieu Labeau, Chloe Clavel

2020-09-23Findings of the Association for Computational Linguistics 2020Text ClassificationEmotion Recognition in ConversationDialogue Act Classification
PaperPDF

Abstract

Sequence labelling tasks like Dialog Act and Emotion/Sentiment identification are a key component of spoken dialog systems. In this work, we propose a new approach to learn generic representations adapted to spoken dialog, which we evaluate on a new benchmark we call Sequence labellIng evaLuatIon benChmark fOr spoken laNguagE benchmark (\texttt{SILICONE}). \texttt{SILICONE} is model-agnostic and contains 10 different datasets of various sizes. We obtain our representations with a hierarchical encoder based on transformer architectures, for which we extend two well-known pre-training objectives. Pre-training is performed on OpenSubtitles: a large corpus of spoken dialog containing over $2.3$ billion of tokens. We demonstrate how hierarchical encoders achieve competitive results with consistently fewer parameters compared to state-of-the-art models and we show their importance for both pre-training and fine-tuning.

Results

TaskDatasetMetricValueModel
DialogueSwitchboard corpusAccuracy79.2Pretrained Hierarchical Transformer
DialogueICSI Meeting Recorder Dialog Act (MRDA) corpusAccuracy92.4Pretrained Hierarchical Transformer
Emotion RecognitionSEMAINEMAE (Arousal)0.16Pretrained Hierarchical Transformer
Emotion RecognitionSEMAINEMAE (Expectancy)0.16Pretrained Hierarchical Transformer
Emotion RecognitionSEMAINEMAE (Power)7.7Pretrained Hierarchical Transformer
Emotion RecognitionSEMAINEMAE (Valence)0.16Pretrained Hierarchical Transformer
Emotion RecognitionMELDWeighted-F161.9Pretrained Hierarchical Transformer
Emotion RecognitionDailyDialogMicro-F160.14Pretrained Hierarchical Transformer
Emotion RecognitionIEMOCAPAccuracy66.05Pretrained Hierarchical Transformer
Emotion RecognitionIEMOCAPWeighted-F165.37Pretrained Hierarchical Transformer
Text ClassificationSILICONE Benchmark1:1 Accuracy71.25Pretrained Hierarchical Transformer
ClassificationSILICONE Benchmark1:1 Accuracy71.25Pretrained Hierarchical Transformer

Related Papers

Long-Short Distance Graph Neural Networks and Improved Curriculum Learning for Emotion Recognition in Conversation2025-07-21Making Language Model a Hierarchical Classifier and Generator2025-07-17Dynamic Parameter Memory: Temporary LoRA-Enhanced LLM for Long-Sequence Emotion Recognition in Conversation2025-07-11GNN-CNN: An Efficient Hybrid Model of Convolutional and Graph Neural Networks for Text Representation2025-07-10The Trilemma of Truth in Large Language Models2025-06-30Robustness of Misinformation Classification Systems to Adversarial Examples Through BeamAttack2025-06-30Perspectives in Play: A Multi-Perspective Approach for More Inclusive NLP Systems2025-06-25Can Generated Images Serve as a Viable Modality for Text-Centric Multimodal Learning?2025-06-21