TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Task Compass: Scaling Multi-task Pre-training with Task Pr...

Task Compass: Scaling Multi-task Pre-training with Task Prefix

Zhuosheng Zhang, Shuohang Wang, Yichong Xu, Yuwei Fang, Wenhao Yu, Yang Liu, Hai Zhao, Chenguang Zhu, Michael Zeng

2022-10-12Question AnsweringSentence CompletionCommon Sense ReasoningSelf-Supervised LearningData AugmentationTransfer LearningMulti-Task Learning
PaperPDFCode(official)

Abstract

Leveraging task-aware annotated data as supervised signals to assist with self-supervised learning on large-scale unlabeled data has become a new trend in pre-training language models. Existing studies show that multi-task learning with large-scale supervised tasks suffers from negative effects across tasks. To tackle the challenge, we propose a task prefix guided multi-task pre-training framework to explore the relationships among tasks. We conduct extensive experiments on 40 datasets, which show that our model can not only serve as the strong foundation backbone for a wide range of tasks but also be feasible as a probing tool for analyzing task relationships. The task relationships reflected by the prefixes align transfer learning performance between tasks. They also suggest directions for data augmentation with complementary tasks, which help our model achieve human-parity results on commonsense reasoning leaderboards. Code is available at https://github.com/cooelf/CompassMTL

Results

TaskDatasetMetricValueModel
Question AnsweringSIQAAccuracy82.2CompassMTL 567M with Tailor
Question AnsweringSIQAAccuracy81.7CompassMTL 567M
Question AnsweringSIQAAccuracy79.6ExDeBERTa 567M
Question AnsweringPIQAAccuracy88.3CompassMTL 567M with Tailor
Question AnsweringPIQAAccuracy87.3CompassMTL 567M
Question AnsweringPIQAAccuracy85.5ExDeBERTa 567M
Common Sense ReasoningWinoGrandeAccuracy90.5CompassMTL 567M with Tailor
Common Sense ReasoningWinoGrandeAccuracy89.6CompassMTL 567M
Common Sense ReasoningWinoGrandeAccuracy87ExDeBERTa 567M
Sentence CompletionHellaSwagAccuracy96.1CompassMTL 567M with Tailor
Sentence CompletionHellaSwagAccuracy95.6CompassMTL 567M
Sentence CompletionHellaSwagAccuracy83.6ExDeBERTa 567M

Related Papers

RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Comparing Apples to Oranges: A Dataset & Analysis of LLM Humour Understanding from Traditional Puns to Topical Jokes2025-07-17A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17