TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Decomposable Attention Model for Natural Language Infere...

A Decomposable Attention Model for Natural Language Inference

Ankur P. Parikh, Oscar Täckström, Dipanjan Das, Jakob Uszkoreit

2016-06-06EMNLP 2016 11Natural Language Inference
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

We propose a simple neural architecture for natural language inference. Our approach uses attention to decompose the problem into subproblems that can be solved separately, thus making it trivially parallelizable. On the Stanford Natural Language Inference (SNLI) dataset, we obtain state-of-the-art results with almost an order of magnitude fewer parameters than previous work and without relying on any word-order information. Adding intra-sentence attention that takes a minimum amount of order into account yields further improvements.

Results

TaskDatasetMetricValueModel
Natural Language InferenceSNLI% Test Accuracy86.8200D decomposable attention feed-forward model with intra-sentence attention
Natural Language InferenceSNLI% Train Accuracy90.5200D decomposable attention feed-forward model with intra-sentence attention
Natural Language InferenceSNLI% Test Accuracy86.8200D decomposable attention model with intra-sentence attention
Natural Language InferenceSNLI% Train Accuracy90.5200D decomposable attention model with intra-sentence attention
Natural Language InferenceSNLI% Test Accuracy86.3200D decomposable attention feed-forward model
Natural Language InferenceSNLI% Train Accuracy89.5200D decomposable attention feed-forward model
Natural Language InferenceSNLI% Test Accuracy86.3200D decomposable attention model
Natural Language InferenceSNLI% Train Accuracy89.5200D decomposable attention model

Related Papers

LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15DS@GT at CheckThat! 2025: Evaluating Context and Tokenization Strategies for Numerical Fact Verification2025-07-08ARAG: Agentic Retrieval Augmented Generation for Personalized Recommendation2025-06-27Thunder-NUBench: A Benchmark for LLMs' Sentence-Level Negation Understanding2025-06-17When Does Meaning Backfire? Investigating the Role of AMRs in NLI2025-06-17Explainable Compliance Detection with Multi-Hop Natural Language Inference on Assurance Case Structure2025-06-10Theorem-of-Thought: A Multi-Agent Framework for Abductive, Deductive, and Inductive Reasoning in Language Models2025-06-08A MISMATCHED Benchmark for Scientific Natural Language Inference2025-06-05