DrugImproverGPT: A Large Language Model for Drug Optimization with Fine-Tuning via Structured Policy Optimization
FOS: Computer and information sciences
Computer Science - Machine Learning
Computer Science - Computation and Language
Quantitative Biology - Biomolecules
Statistics - Machine Learning
FOS: Biological sciences
Biomolecules (q-bio.BM)
Machine Learning (stat.ML)
Computation and Language (cs.CL)
Machine Learning (cs.LG)
DOI:
10.48550/arxiv.2502.07237
Publication Date:
2025-02-10
AUTHORS (7)
ABSTRACT
Finetuning a Large Language Model (LLM) is crucial for generating results towards specific objectives. This research delves into the realm of drug optimization and introduce novel reinforcement learning algorithm to finetune LLM-based generative model, enhancing original across target objectives, while retains beneficial chemical properties drug. work comprised two primary components: (1) DrugImprover: A framework tailored improving robustness efficiency in optimization. It includes LLM designed Structured Policy Optimization (SPO) algorithm, which theoretically grounded. offers unique perspective fine-tuning model by aligning improvement generated molecule with input under desired (2) dataset 1 million compounds, each OEDOCK docking scores on 5 human proteins associated cancer cells 24 binding sites from SARS-CoV-2 virus. We conduct comprehensive evaluation SPO demonstrate its effectiveness properties. Our code will be publicly available at: https://github.com/xuefeng-cs/DrugImproverGPT.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....