TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MolFM: A Multimodal Molecular Foundation Model

MolFM: A Multimodal Molecular Foundation Model

Yizhen Luo, Kai Yang, Massimo Hong, Xing Yi Liu, Zaiqing Nie

2023-06-06Cross-Modal RetrievalKnowledge GraphsRepresentation LearningText-based de novo Molecule GenerationRetrievalMolecule Captioning
PaperPDFCodeCode(official)

Abstract

Molecular knowledge resides within three different modalities of information sources: molecular structures, biomedical documents, and knowledge bases. Effective incorporation of molecular knowledge from these modalities holds paramount significance in facilitating biomedical research. However, existing multimodal molecular foundation models exhibit limitations in capturing intricate connections between molecular structures and texts, and more importantly, none of them attempt to leverage a wealth of molecular expertise derived from knowledge graphs. In this study, we introduce MolFM, a multimodal molecular foundation model designed to facilitate joint representation learning from molecular structures, biomedical texts, and knowledge graphs. We propose cross-modal attention between atoms of molecular structures, neighbors of molecule entities and semantically related texts to facilitate cross-modal comprehension. We provide theoretical analysis that our cross-modal pre-training captures local and global molecular knowledge by minimizing the distance in the feature space between different modalities of the same molecule, as well as molecules sharing similar structures or functions. MolFM achieves state-of-the-art performance on various downstream tasks. On cross-modal retrieval, MolFM outperforms existing models with 12.13% and 5.04% absolute gains under the zero-shot and fine-tuning settings, respectively. Furthermore, qualitative analysis showcases MolFM's implicit ability to provide grounding from molecular substructures and knowledge graphs. Code and models are available on https://github.com/BioFM/OpenBioMed.

Results

TaskDatasetMetricValueModel
Drug DiscoveryChEBI-20BLEU82.2MolFM-Base
Drug DiscoveryChEBI-20Exact Match21MolFM-Base
Drug DiscoveryChEBI-20Levenshtein19.445MolFM-Base
Drug DiscoveryChEBI-20MACCS FTS85.4MolFM-Base
Drug DiscoveryChEBI-20Morgan FTS75.8MolFM-Base
Drug DiscoveryChEBI-20Parameter Count296200000MolFM-Base
Drug DiscoveryChEBI-20RDK FTS69.7MolFM-Base
Drug DiscoveryChEBI-20Text2Mol58.3MolFM-Base
Drug DiscoveryChEBI-20Validity89.2MolFM-Base
Drug DiscoveryChEBI-20BLEU80.3MolFM-Small
Drug DiscoveryChEBI-20Exact Match16.9MolFM-Small
Drug DiscoveryChEBI-20Levenshtein20.868MolFM-Small
Drug DiscoveryChEBI-20MACCS FTS83.4MolFM-Small
Drug DiscoveryChEBI-20Morgan FTS72.1MolFM-Small
Drug DiscoveryChEBI-20Parameter Count13620000MolFM-Small
Drug DiscoveryChEBI-20RDK FTS66.2MolFM-Small
Drug DiscoveryChEBI-20Text2Mol57.3MolFM-Small
Drug DiscoveryChEBI-20Validity85.9MolFM-Small
Molecule CaptioningChEBI-20BLEU-258.5MolFM-Base
Molecule CaptioningChEBI-20BLEU-449.8MolFM-Base
Molecule CaptioningChEBI-20METEOR60.7MolFM-Base
Molecule CaptioningChEBI-20ROUGE-165.3MolFM-Base
Molecule CaptioningChEBI-20ROUGE-250.8MolFM-Base
Molecule CaptioningChEBI-20ROUGE-L59.4MolFM-Base
Molecule CaptioningChEBI-20Text2Mol57.6MolFM-Base
Molecule CaptioningChEBI-20BLEU-254.2MolFM-Small
Molecule CaptioningChEBI-20BLEU-445.2MolFM-Small
Molecule CaptioningChEBI-20METEOR56.4MolFM-Small
Molecule CaptioningChEBI-20ROUGE-162.3MolFM-Small
Molecule CaptioningChEBI-20ROUGE-246.9MolFM-Small
Molecule CaptioningChEBI-20ROUGE-L56.2MolFM-Small
Molecule CaptioningChEBI-20Text2Mol55.7MolFM-Small
Text-based de novo Molecule GenerationChEBI-20BLEU82.2MolFM-Base
Text-based de novo Molecule GenerationChEBI-20Exact Match21MolFM-Base
Text-based de novo Molecule GenerationChEBI-20Levenshtein19.445MolFM-Base
Text-based de novo Molecule GenerationChEBI-20MACCS FTS85.4MolFM-Base
Text-based de novo Molecule GenerationChEBI-20Morgan FTS75.8MolFM-Base
Text-based de novo Molecule GenerationChEBI-20Parameter Count296200000MolFM-Base
Text-based de novo Molecule GenerationChEBI-20RDK FTS69.7MolFM-Base
Text-based de novo Molecule GenerationChEBI-20Text2Mol58.3MolFM-Base
Text-based de novo Molecule GenerationChEBI-20Validity89.2MolFM-Base
Text-based de novo Molecule GenerationChEBI-20BLEU80.3MolFM-Small
Text-based de novo Molecule GenerationChEBI-20Exact Match16.9MolFM-Small
Text-based de novo Molecule GenerationChEBI-20Levenshtein20.868MolFM-Small
Text-based de novo Molecule GenerationChEBI-20MACCS FTS83.4MolFM-Small
Text-based de novo Molecule GenerationChEBI-20Morgan FTS72.1MolFM-Small
Text-based de novo Molecule GenerationChEBI-20Parameter Count13620000MolFM-Small
Text-based de novo Molecule GenerationChEBI-20RDK FTS66.2MolFM-Small
Text-based de novo Molecule GenerationChEBI-20Text2Mol57.3MolFM-Small
Text-based de novo Molecule GenerationChEBI-20Validity85.9MolFM-Small

Related Papers

Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20SMART: Relation-Aware Learning of Geometric Representations for Knowledge Graphs2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17MCoT-RE: Multi-Faceted Chain-of-Thought and Re-Ranking for Training-Free Zero-Shot Composed Image Retrieval2025-07-17