Attention Boosted Sequential Inference Model
Guanyu Li, Pengfei Zhang, Caiyan Jia
2018-12-05Natural Language Inference
Abstract
Attention mechanism has been proven effective on natural language processing. This paper proposes an attention boosted natural language inference model named aESIM by adding word attention and adaptive direction-oriented attention mechanisms to the traditional Bi-LSTM layer of natural language inference models, e.g. ESIM. This makes the inference model aESIM has the ability to effectively learn the representation of words and model the local subsentential inference between pairs of premise and hypothesis. The empirical studies on the SNLI, MultiNLI and Quora benchmarks manifest that aESIM is superior to the original ESIM model.
Results
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Natural Language Inference | Quora Question Pairs | Accuracy | 88.01 | aESIM |
| Natural Language Inference | SNLI | % Test Accuracy | 88.1 | aESIM |
| Natural Language Inference | MultiNLI | Matched | 73.9 | aESIM |
| Natural Language Inference | MultiNLI | Mismatched | 73.9 | aESIM |
Related Papers
LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15DS@GT at CheckThat! 2025: Evaluating Context and Tokenization Strategies for Numerical Fact Verification2025-07-08ARAG: Agentic Retrieval Augmented Generation for Personalized Recommendation2025-06-27Thunder-NUBench: A Benchmark for LLMs' Sentence-Level Negation Understanding2025-06-17When Does Meaning Backfire? Investigating the Role of AMRs in NLI2025-06-17Explainable Compliance Detection with Multi-Hop Natural Language Inference on Assurance Case Structure2025-06-10Theorem-of-Thought: A Multi-Agent Framework for Abductive, Deductive, and Inductive Reasoning in Language Models2025-06-08A MISMATCHED Benchmark for Scientific Natural Language Inference2025-06-05