TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Phraseformer: Multimodal Key-phrase Extraction using Trans...

Phraseformer: Multimodal Key-phrase Extraction using Transformer and Graph Embedding

Narjes Nikzad-Khasmakhi, Mohammad-Reza Feizi-Derakhshi, Meysam Asgari-Chenaghlu, Mohammad-Ali Balafar, Ali-Reza Feizi-Derakhshi, Taymaz Rahkar-Farshi, Majid Ramezani, Zoleikha Jahanbakhsh-Nagadeh, Elnaz Zafarani-Moattar, Mehrdad Ranjbar-Khadivi

2021-06-09Graph EmbeddingKeyword Extraction
PaperPDF

Abstract

Background: Keyword extraction is a popular research topic in the field of natural language processing. Keywords are terms that describe the most relevant information in a document. The main problem that researchers are facing is how to efficiently and accurately extract the core keywords from a document. However, previous keyword extraction approaches have utilized the text and graph features, there is the lack of models that can properly learn and combine these features in a best way. Methods: In this paper, we develop a multimodal Key-phrase extraction approach, namely Phraseformer, using transformer and graph embedding techniques. In Phraseformer, each keyword candidate is presented by a vector which is the concatenation of the text and structure learning representations. Phraseformer takes the advantages of recent researches such as BERT and ExEm to preserve both representations. Also, the Phraseformer treats the key-phrase extraction task as a sequence labeling problem solved using classification task. Results: We analyze the performance of Phraseformer on three datasets including Inspec, SemEval2010 and SemEval 2017 by F1-score. Also, we investigate the performance of different classifiers on Phraseformer method over Inspec dataset. Experimental results demonstrate the effectiveness of Phraseformer method over the three datasets used. Additionally, the Random Forest classifier gain the highest F1-score among all classifiers. Conclusions: Due to the fact that the combination of BERT and ExEm is more meaningful and can better represent the semantic of words. Hence, Phraseformer significantly outperforms single-modality methods.

Results

TaskDatasetMetricValueModel
Keyword ExtractionSemEval-2017 Task-10F1 score67.13Phraseformer(BERT, ExEm(ft))
Keyword ExtractionSemEval-2017 Task-10F1 score66.96Phraseformer(BERT, ExEm(w2v))
Keyword ExtractionSemEval-2017 Task-10F1 score65.94Phraseformer(BERT, Node2vec)
Keyword ExtractionSemEval-2017 Task-10F1 score65.7Phraseformer(BERT, DeepWalk)
Keyword ExtractionInspecF1 score69.87Phraseformer(BERT, ExEm(ft))
Keyword ExtractionInspecF1 score69.7Phraseformer(BERT, ExEm(w2v))
Keyword ExtractionInspecF1 score68.68Phraseformer(BERT, Node2vec)
Keyword ExtractionInspecF1 score68.44Phraseformer(BERT, DeepWalk)
Keyword ExtractionSemEval 2010 Task 8F1 score48.65Phraseformer(BERT, ExEm(ft))
Keyword ExtractionSemEval 2010 Task 8F1 score48.48Phraseformer(BERT, ExEm(w2v))
Keyword ExtractionSemEval 2010 Task 8F1 score47.46Phraseformer(BERT, Node2vec)
Keyword ExtractionSemEval 2010 Task 8F1 score47.22Phraseformer(BERT, DeepWalk)

Related Papers

SMART: Relation-Aware Learning of Geometric Representations for Knowledge Graphs2025-07-17Metapath-based Hyperbolic Contrastive Learning for Heterogeneous Graph Embedding2025-06-20Cost-Efficient Serving of LLM Agents via Test-Time Plan Caching2025-06-17ETT-CKGE: Efficient Task-driven Tokens for Continual Knowledge Graph Embedding2025-06-09Urania: Differentially Private Insights into AI Use2025-06-05Efficient Identity and Position Graph Embedding via Spectral-Based Random Feature Aggregation2025-05-27Predicate-Conditional Conformalized Answer Sets for Knowledge Graph Embeddings2025-05-22Lightweight Spatio-Temporal Attention Network with Graph Embedding and Rotational Position Encoding for Traffic Forecasting2025-05-17