TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Induction Networks for Few-Shot Text Classification

Induction Networks for Few-Shot Text Classification

Ruiying Geng, Binhua Li, Yongbin Li, Xiaodan Zhu, Ping Jian, Jian Sun

2019-02-27IJCNLP 2019 11Text ClassificationMeta-Learningintent-classificationSentiment AnalysisSentiment ClassificationFew-Shot Text ClassificationGeneral ClassificationIntent Classification
PaperPDFCodeCodeCodeCodeCode

Abstract

Text classification tends to struggle when data is deficient or when it needs to adapt to unseen classes. In such challenging scenarios, recent studies have used meta-learning to simulate the few-shot task, in which new queries are compared to a small support set at the sample-wise level. However, this sample-wise comparison may be severely disturbed by the various expressions in the same class. Therefore, we should be able to learn a general representation of each class in the support set and then compare it to new queries. In this paper, we propose a novel Induction Network to learn such a generalized class-wise representation, by innovatively leveraging the dynamic routing algorithm in meta-learning. In this way, we find the model is able to induce and generalize better. We evaluate the proposed model on a well-studied sentiment classification dataset (English) and a real-world dialogue intent classification dataset (Chinese). Experiment results show that on both datasets, the proposed model significantly outperforms the existing state-of-the-art approaches, proving the effectiveness of class-wise generalization in few-shot text classification.

Results

TaskDatasetMetricValueModel
Text ClassificationODIC 5-way (5-shot)Accuracy87.16Induction Networks
Text ClassificationODIC 5-way (10-shot)Accuracy88.49Induction Networks
Text ClassificationODIC 10-way (5-shot)Accuracy78.27Induction Networks
Text ClassificationODIC 10-way (10-shot)Accuracy81.64Induction Networks
Few-Shot Text ClassificationODIC 5-way (5-shot)Accuracy87.16Induction Networks
Few-Shot Text ClassificationODIC 5-way (10-shot)Accuracy88.49Induction Networks
Few-Shot Text ClassificationODIC 10-way (5-shot)Accuracy78.27Induction Networks
Few-Shot Text ClassificationODIC 10-way (10-shot)Accuracy81.64Induction Networks
ClassificationODIC 5-way (5-shot)Accuracy87.16Induction Networks
ClassificationODIC 5-way (10-shot)Accuracy88.49Induction Networks
ClassificationODIC 10-way (5-shot)Accuracy78.27Induction Networks
ClassificationODIC 10-way (10-shot)Accuracy81.64Induction Networks

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17AdaptiSent: Context-Aware Adaptive Attention for Multimodal Aspect-Based Sentiment Analysis2025-07-17Are encoders able to learn landmarkers for warm-starting of Hyperparameter Optimization?2025-07-16Imbalanced Regression Pipeline Recommendation2025-07-16CLID-MU: Cross-Layer Information Divergence Based Meta Update Strategy for Learning with Noisy Labels2025-07-16Mixture of Experts in Large Language Models2025-07-15AI Wizards at CheckThat! 2025: Enhancing Transformer-Based Embeddings with Sentiment for Subjectivity Detection in News Articles2025-07-15DCR: Quantifying Data Contamination in LLMs Evaluation2025-07-15