TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Guess the Instruction! Flipped Learning Makes Language Mod...

Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners

Seonghyeon Ye, Doyoung Kim, Joel Jang, Joongbo Shin, Minjoon Seo

2022-10-06Question AnsweringSentence CompletionCoreference ResolutionNatural Language InferenceCommon Sense ReasoningNatural Language Inference (Zero-Shot)Word Sense DisambiguationLanguage Modelling
PaperPDFCode(official)

Abstract

Meta-training, which fine-tunes the language model (LM) on various downstream tasks by maximizing the likelihood of the target label given the task instruction and input instance, has improved the zero-shot task generalization performance. However, meta-trained LMs still struggle to generalize to challenging tasks containing novel labels unseen during meta-training. In this paper, we propose Flipped Learning, an alternative method of meta-training which trains the LM to generate the task instruction given the input instance and label. During inference, the LM trained with Flipped Learning, referred to as Flipped, selects the label option that is most likely to generate the task instruction. On 14 tasks of the BIG-bench benchmark, the 11B-sized Flipped outperforms zero-shot T0-11B and even a 16 times larger 3-shot GPT-3 (175B) on average by 8.4% and 9.7% points, respectively. Flipped gives particularly large improvements on tasks with unseen labels, outperforming T0-11B by up to +20% average F1 score. This indicates that the strong task generalization of Flipped comes from improved generalization to novel labels. We release our code at https://github.com/seonghyeonye/Flipped-Learning.

Results

TaskDatasetMetricValueModel
Question AnsweringCOPAAccuracy89.88Flipped-3B
Question AnsweringStoryClozeAccuracy95.88Flipped-3B
Common Sense ReasoningWinoGrandeAccuracy58.56Flipped-3B
Word Sense DisambiguationWords in ContextAccuracy50.42Flipped-3B
Natural Language InferenceANLI testA139.99Flipped-3B
Natural Language InferenceANLI testA237.05Flipped-3B
Natural Language InferenceANLI testA337.73Flipped-3B
Natural Language InferenceRTEAccuracy71.05Flipped-3B
Coreference ResolutionWinograd Schema ChallengeAccuracy58.37Flipped-3B
Sentence CompletionHellaSwagAccuracy41.6Flipped-3B

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Comparing Apples to Oranges: A Dataset & Analysis of LLM Humour Understanding from Traditional Puns to Topical Jokes2025-07-17Making Language Model a Hierarchical Classifier and Generator2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17