TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Hyperbolic Representation Learning for Fast and Efficient ...

Hyperbolic Representation Learning for Fast and Efficient Neural Question Answering

Yi Tay, Luu Anh Tuan, Siu Cheung Hui

2017-07-25Feature EngineeringQuestion AnsweringRepresentation LearningRetrieval
PaperPDFCode

Abstract

The dominant neural architectures in question answer retrieval are based on recurrent or convolutional encoders configured with complex word matching layers. Given that recent architectural innovations are mostly new word interaction layers or attention-based matching mechanisms, it seems to be a well-established fact that these components are mandatory for good performance. Unfortunately, the memory and computation cost incurred by these complex mechanisms are undesirable for practical applications. As such, this paper tackles the question of whether it is possible to achieve competitive performance with simple neural architectures. We propose a simple but novel deep learning architecture for fast and efficient question-answer ranking and retrieval. More specifically, our proposed model, \textsc{HyperQA}, is a parameter efficient neural network that outperforms other parameter intensive models such as Attentive Pooling BiLSTMs and Multi-Perspective CNNs on multiple QA benchmarks. The novelty behind \textsc{HyperQA} is a pairwise ranking objective that models the relationship between question and answer embeddings in Hyperbolic space instead of Euclidean space. This empowers our model with a self-organizing ability and enables automatic discovery of latent hierarchies while learning embeddings of questions and answers. Our model requires no feature engineering, no similarity matrix matching, no complicated attention mechanisms nor over-parameterized layers and yet outperforms and remains competitive to many models that have these functionalities on multiple benchmarks.

Results

TaskDatasetMetricValueModel
Question AnsweringSemEvalCQAMAP0.795HyperQA
Question AnsweringSemEvalCQAP@10.809HyperQA
Question AnsweringYahooCQAMRR0.801HyperQA
Question AnsweringYahooCQAP@10.683HyperQA
Question AnsweringYahooCQAMRR0.669LSTM
Question AnsweringYahooCQAP@10.465LSTM
Question AnsweringYahooCQAMRR0.632CNN
Question AnsweringYahooCQAP@10.413CNN
Question AnsweringTrecQAMAP0.77HyperQA
Question AnsweringTrecQAMRR0.825HyperQA
Question AnsweringWikiQAMAP0.712HyperQA
Question AnsweringWikiQAMRR0.727HyperQA

Related Papers

Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17