TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Molecular Multimodal Foundation Model Associating Molecu...

A Molecular Multimodal Foundation Model Associating Molecule Graphs with Natural Language

Bing Su, Dazhao Du, Zhao Yang, Yujie Zhou, Jiangmeng Li, Anyi Rao, Hao Sun, Zhiwu Lu, Ji-Rong Wen

2022-09-12Molecular Property PredictionCross-Modal RetrievalContrastive LearningRetrievalMolecule Captioning
PaperPDFCodeCode(official)CodeCode

Abstract

Although artificial intelligence (AI) has made significant progress in understanding molecules in a wide range of fields, existing models generally acquire the single cognitive ability from the single molecular modality. Since the hierarchy of molecular knowledge is profound, even humans learn from different modalities including both intuitive diagrams and professional texts to assist their understanding. Inspired by this, we propose a molecular multimodal foundation model which is pretrained from molecular graphs and their semantically related textual data (crawled from published Scientific Citation Index papers) via contrastive learning. This AI model represents a critical attempt that directly bridges molecular graphs and natural language. Importantly, through capturing the specific and complementary information of the two modalities, our proposed model can better grasp molecular expertise. Experimental results show that our model not only exhibits promising performance in cross-modal tasks such as cross-modal retrieval and molecule caption, but also enhances molecular property prediction and possesses capability to generate meaningful molecular graphs from natural language descriptions. We believe that our model would have a broad impact on AI-empowered fields across disciplines such as biology, chemistry, materials, environment, and medicine, among others.

Results

TaskDatasetMetricValueModel
Molecule CaptioningChEBI-20BLEU-259.9MoMu+MolT5-Large
Molecule CaptioningChEBI-20BLEU-451.5MoMu+MolT5-Large
Molecule CaptioningChEBI-20METEOR59.7MoMu+MolT5-Large
Molecule CaptioningChEBI-20Text2Mol58.2MoMu+MolT5-Large
Molecule CaptioningChEBI-20BLEU-254.9MoMu+MolT5-Base
Molecule CaptioningChEBI-20BLEU-446.2MoMu+MolT5-Base
Molecule CaptioningChEBI-20METEOR57.6MoMu+MolT5-Base
Molecule CaptioningChEBI-20Text2Mol55.8MoMu+MolT5-Base
Molecule CaptioningChEBI-20BLEU-253.2MoMu+MolT5-Small
Molecule CaptioningChEBI-20BLEU-444.5MoMu+MolT5-Small
Molecule CaptioningChEBI-20METEOR55.7MoMu+MolT5-Small
Molecule CaptioningChEBI-20Text2Mol55.3MoMu+MolT5-Small

Related Papers

SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17MCoT-RE: Multi-Faceted Chain-of-Thought and Re-Ranking for Training-Free Zero-Shot Composed Image Retrieval2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16