TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/TRANS-BLSTM: Transformer with Bidirectional LSTM for Langu...

TRANS-BLSTM: Transformer with Bidirectional LSTM for Language Understanding

Zhiheng Huang, Peng Xu, Davis Liang, Ajay Mishra, Bing Xiang

2020-03-16Text ClassificationMachine TranslationQuestion AnsweringParaphrase IdentificationNatural Language InferenceTranslationSentence Classification
PaperPDF

Abstract

Bidirectional Encoder Representations from Transformers (BERT) has recently achieved state-of-the-art performance on a broad range of NLP tasks including sentence classification, machine translation, and question answering. The BERT model architecture is derived primarily from the transformer. Prior to the transformer era, bidirectional Long Short-Term Memory (BLSTM) has been the dominant modeling architecture for neural machine translation and question answering. In this paper, we investigate how these two modeling techniques can be combined to create a more powerful model architecture. We propose a new architecture denoted as Transformer with BLSTM (TRANS-BLSTM) which has a BLSTM layer integrated to each transformer block, leading to a joint modeling framework for transformer and BLSTM. We show that TRANS-BLSTM models consistently lead to improvements in accuracy compared to BERT baselines in GLUE and SQuAD 1.1 experiments. Our TRANS-BLSTM model obtains an F1 score of 94.01% on the SQuAD 1.1 development dataset, which is comparable to the state-of-the-art result.

Results

TaskDatasetMetricValueModel
Semantic Textual SimilarityQuora Question PairsAccuracy88.28TRANS-BLSTM
Paraphrase IdentificationQuora Question PairsAccuracy88.28TRANS-BLSTM
Text ClassificationGLUE SST2Accuracy94.38TRANS-BLSTM
Text ClassificationGLUE RTEAccuracy79.78TRANS-BLSTM
Text ClassificationGLUE MRPCAccuracy90.45TRANS-BLSTM
ClassificationGLUE SST2Accuracy94.38TRANS-BLSTM
ClassificationGLUE RTEAccuracy79.78TRANS-BLSTM
ClassificationGLUE MRPCAccuracy90.45TRANS-BLSTM

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16Is This Just Fantasy? Language Model Representations Reflect Human Judgments of Event Plausibility2025-07-16