TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Self-feedback Knowledge Elicitation Approach for Chemica...

A Self-feedback Knowledge Elicitation Approach for Chemical Reaction Predictions

PengFei Liu, Jun Tao, Zhixiang Ren

2024-04-15Drug DiscoveryRetrosynthesisLarge Language ModelChemical Reaction PredictionLanguage Modelling
PaperPDFCode(official)

Abstract

The task of chemical reaction predictions (CRPs) plays a pivotal role in advancing drug discovery and material science. However, its effectiveness is constrained by the vast and uncertain chemical reaction space and challenges in capturing reaction selectivity, particularly due to existing methods' limitations in exploiting the data's inherent knowledge. To address these challenges, we introduce a data-curated self-feedback knowledge elicitation approach. This method starts from iterative optimization of molecular representations and facilitates the extraction of knowledge on chemical reaction types (RTs). Then, we employ adaptive prompt learning to infuse the prior knowledge into the large language model (LLM). As a result, we achieve significant enhancements: a 14.2% increase in retrosynthesis prediction accuracy, a 74.2% rise in reagent prediction accuracy, and an expansion in the model's capability for handling multi-task chemical reactions. This research offers a novel paradigm for knowledge elicitation in scientific research and showcases the untapped potential of LLMs in CRPs.

Results

TaskDatasetMetricValueModel
Chemical Reaction PredictionMol-InstructionExact0.674SLM4CRP
Chemical Reaction PredictionMol-InstructionMETEOR0.901SLM4CRP
Chemical Reaction PredictionMol-InstructionMorgan FTS0.854SLM4CRP
Chemical Reaction PredictionMol-InstructionValidity0.998SLM4CRP
Forward reaction predictionMol-InstructionExact0.945SLM4CRP
Forward reaction predictionMol-InstructionMETEOR0.993SLM4CRP
Forward reaction predictionMol-InstructionMorgan FTS0.986SLM4CRP
Forward reaction predictionMol-InstructionValidity0.997SLM4CRP
Reagent PredictionMol-InstructionExact0.284SLM4CRP
Reagent PredictionMol-InstructionMETEOR0.744SLM4CRP
Reagent PredictionMol-InstructionMorgan FTS0.649SLM4CRP
Reagent PredictionMol-InstructionValidity1SLM4CRP
RetrosynthesisMol-InstructionExact0.757SLM4CRP
RetrosynthesisMol-InstructionMETEOR0.95SLM4CRP
RetrosynthesisMol-InstructionMorgan FTS0.905SLM4CRP
RetrosynthesisMol-InstructionValidity0.994SLM4CRP

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21DENSE: Longitudinal Progress Note Generation with Temporal Modeling of Heterogeneous Clinical Notes Across Hospital Visits2025-07-18GeoReg: Weight-Constrained Few-Shot Regression for Socio-Economic Estimation using LLM2025-07-17The Generative Energy Arena (GEA): Incorporating Energy Awareness in Large Language Model (LLM) Human Evaluations2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Rethinking the Embodied Gap in Vision-and-Language Navigation: A Holistic Study of Physical and Visual Disparities2025-07-17Making Language Model a Hierarchical Classifier and Generator2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17