TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Attentive Pooling Networks

Attentive Pooling Networks

Cicero dos Santos, Ming Tan, Bing Xiang, Bo-Wen Zhou

2016-02-11Question AnsweringRepresentation LearningAnswer Selection
PaperPDFCodeCodeCode

Abstract

In this work, we propose Attentive Pooling (AP), a two-way attention mechanism for discriminative model training. In the context of pair-wise ranking or classification with neural networks, AP enables the pooling layer to be aware of the current input pair, in a way that information from the two input items can directly influence the computation of each other's representations. Along with such representations of the paired inputs, AP jointly learns a similarity measure over projected segments (e.g. trigrams) of the pair, and subsequently, derives the corresponding attention vector for each input to guide the pooling. Our two-way attention mechanism is a general framework independent of the underlying representation learning, and it has been applied to both convolutional neural networks (CNNs) and recurrent neural networks (RNNs) in our studies. The empirical results, from three very different benchmark tasks of question answering/answer selection, demonstrate that our proposed models outperform a variety of strong baselines and achieve state-of-the-art performance in all the benchmarks.

Results

TaskDatasetMetricValueModel
Question AnsweringSemEvalCQAMAP0.771AP-CNN
Question AnsweringSemEvalCQAP@10.755AP-CNN
Question AnsweringYahooCQAMRR0.731AP-BiLSTM
Question AnsweringYahooCQAP@10.568AP-BiLSTM
Question AnsweringYahooCQAMRR0.726AP-CNN
Question AnsweringYahooCQAP@10.56AP-CNN
Question AnsweringWikiQAMAP0.6886AP-CNN
Question AnsweringWikiQAMRR0.6957AP-CNN

Related Papers

Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16