TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/T5

T5

Natural Language ProcessingIntroduced 2000708 papers
Source Paper

Description

T5, or Text-to-Text Transfer Transformer, is a Transformer based architecture that uses a text-to-text approach. Every task – including translation, question answering, and classification – is cast as feeding the model text as input and training it to generate some target text. This allows for the use of the same model, loss function, hyperparameters, etc. across our diverse set of tasks. The changes compared to BERT include:

  • adding a causal decoder to the bidirectional architecture.
  • replacing the fill-in-the-blank cloze task with a mix of alternative pre-training tasks.

Papers Using This Method

LiLM-RDB-SFC: Lightweight Language Model with Relational Database-Guided DRL for Optimized SFC Provisioning2025-07-15Chat-Ghosting: A Comparative Study of Methods for Auto-Completion in Dialog Systems2025-07-08I Know Which LLM Wrote Your Code Last Summer: LLM generated Code Stylometry for Authorship Attribution2025-06-18Fretting-Transformer: Encoder-Decoder Model for MIDI to Tablature Transcription2025-06-17A Comprehensive Study of Decoder-Only LLMs for Text-to-Image Generation2025-06-09A Multi-Dataset Evaluation of Models for Automated Vulnerability Repair2025-06-05Decom-Renorm-Merge: Model Merging on the Right Space Improves Multitasking2025-05-29ShIOEnv: A CLI Behavior-Capturing Environment Enabling Grammar-Guided Command Synthesis for Dataset Curation2025-05-23LogiCase: Effective Test Case Generation from Logical Description in Competitive Programming2025-05-21EEG-to-Text Translation: A Model for Deciphering Human Brain Activity2025-05-20Masking in Multi-hop QA: An Analysis of How Language Models Perform with Context Permutation2025-05-16Multilingual Machine Translation with Quantum Encoder Decoder Attention-based Convolutional Variational Circuits2025-05-14Performance Evaluation of Large Language Models in Bangla Consumer Health Query Summarization2025-05-08GASCADE: Grouped Summarization of Adverse Drug Event for Enhanced Cancer Pharmacovigilance2025-05-07A review of DNA restriction-free overlapping sequence cloning techniques for synthetic biology2025-05-06JaccDiv: A Metric and Benchmark for Quantifying Diversity of Generated Marketing Text in the Music Industry2025-04-29Large Language Models are Qualified Benchmark Builders: Rebuilding Pre-Training Datasets for Advancing Code Intelligence Tasks2025-04-28Sigma: A dataset for text-to-code semantic parsing with statistical analysis2025-04-05Advancing Sentiment Analysis in Tamil-English Code-Mixed Texts: Challenges and Transformer-Based Solutions2025-03-30Enhancing Knowledge Graph Completion with Entity Neighborhood and Relation Context2025-03-29