Discovering Photoswitchable Molecules for Drug Delivery with Large Language Models and Chemist Instruction Training

被引:0
|
作者
Hu, Junjie [1 ]
Wu, Peng [2 ]
Li, Yulin [3 ]
Li, Qi [1 ]
Wang, Shiyi [1 ]
Liu, Yang [4 ]
Qian, Kun [5 ]
Yang, Guang [1 ,6 ,7 ,8 ]
机构
[1] Imperial Coll London, Bioengn Dept & Imperial X, London W12 7SL, England
[2] Ningxia Univ, Sch Chem & Chem Engn, Yinchuan 750014, Peoples R China
[3] Chinese Univ Hong Kong, Dept Math, Shatin, Hong Kong, Peoples R China
[4] Shanxi Med Univ, Shanxi Bethune Hosp, Tongji Shanxi Hosp, Shanxi Acad Med Sci,3 Hosp, Taiyuan 030032, Peoples R China
[5] Fudan Univ, Zhongshan Hosp, Dept Informat & Intelligence Dev, 180 Fenglin Rd, Shanghai 200032, Peoples R China
[6] Imperial Coll London, Natl Heart & Lung Inst, London SW7 2AZ, England
[7] Royal Brompton Hosp, Cardiovasc Res Ctr, London SW3 6NP, England
[8] Kings Coll London, Sch Biomed Engn & Imaging Sci, London WC2R 2LS, England
关键词
drug delivery; photoresponsive molecules; quantum chemistry; language model; RLHF; ABSORPTION; ENERGIES;
D O I
10.3390/ph17101300
中图分类号
R914 [药物化学];
学科分类号
100701 ;
摘要
Background: As large language models continue to expand in size and diversity, their substantial potential and the relevance of their applications are increasingly being acknowledged. The rapid advancement of these models also holds profound implications for the long-term design of stimulus-responsive materials used in drug delivery. Methods: The large model used Hugging Face's Transformers package with BigBird, Gemma, and GPT NeoX architectures. Pre-training used the PubChem dataset, and fine-tuning used QM7b. Chemist instruction training was based on Direct Preference Optimization. Drug Likeness, Synthetic Accessibility, and PageRank Scores were used to filter molecules. All computational chemistry simulations were performed using ORCA and Time-Dependent Density-Functional Theory. Results: To optimize large models for extensive dataset processing and comprehensive learning akin to a chemist's intuition, the integration of deeper chemical insights is imperative. Our study initially compared the performance of BigBird, Gemma, GPT NeoX, and others, specifically focusing on the design of photoresponsive drug delivery molecules. We gathered excitation energy data through computational chemistry tools and further investigated light-driven isomerization reactions as a critical mechanism in drug delivery. Additionally, we explored the effectiveness of incorporating human feedback into reinforcement learning to imbue large models with chemical intuition, enhancing their understanding of relationships involving -N=N- groups in the photoisomerization transitions of photoresponsive molecules. Conclusions: We implemented an efficient design process based on structural knowledge and data, driven by large language model technology, to obtain a candidate dataset of specific photoswitchable molecules. However, the lack of specialized domain datasets remains a challenge for maximizing model performance.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Discovering prerequisite relations using large language models
    Aytekin, Mehmet Cem
    Saygin, Yucel
    INTERACTIVE LEARNING ENVIRONMENTS, 2024,
  • [2] OCTOPACK: INSTRUCTION TUNING CODE LARGE LANGUAGE MODELS
    Muennighoff, Niklas
    Liu, Qian
    Zebaze, Armel
    Zheng, Qinkai
    Hui, Binyuan
    Zhuo, Terry Yue
    Singh, Swayam
    Tang, Xiangru
    von Werra, Leandro
    Longpre, Shayne
    arXiv, 2023,
  • [3] GraphGPT: Graph Instruction Tuning for Large Language Models
    Tang, Jiabin
    Yang, Yuhao
    Wei, Wei
    Shi, Lei
    Su, Lixin
    Cheng, Suqi
    Yin, Dawei
    Huang, Chao
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 491 - 500
  • [4] Emerging opportunities of using large language models for translation between drug molecules and indications
    Oniani, David
    Hilsman, Jordan
    Zang, Chengxi
    Wang, Junmei
    Cai, Lianjin
    Zawala, Jan
    Wang, Yanshan
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [5] Can large language models understand molecules?
    Sadeghi, Shaghayegh
    Bui, Alan
    Forooghi, Ali
    Lu, Jianguo
    Ngom, Alioune
    BMC BIOINFORMATICS, 2024, 25 (01):
  • [6] BioInstruct: instruction tuning of large language models for biomedical natural language processing
    Tran, Hieu
    Yang, Zhichao
    Yao, Zonghai
    Yu, Hong
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 1821 - 1832
  • [7] Phased Instruction Fine-Tuning for Large Language Models
    Pang, Wei
    Zhou, Chuan
    Zhou, Xiao-Hua
    Wang, Xiaojie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5735 - 5748
  • [8] Hybrid Alignment Training for Large Language Models
    Wang, Chenglong
    Zhou, Hang
    Chang, Kaiyan
    Li, Bei
    Mu, Yongyu
    Xiao, Tong
    Liu, Tongran
    Zhu, Jingbo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 11389 - 11403
  • [9] Navigation Instruction Generation with BEV Perception and Large Language Models
    Fan, Sheng
    Liu, Rui
    Wang, Wenguan
    Yang, Yi
    COMPUTER VISION-ECCV 2024, PT XXII, 2025, 15080 : 368 - 387
  • [10] Smaller Language Models are capable of selecting Instruction-Tuning Training Data for Larger Language Models
    Mekala, Dheeraj
    Nguyen, Alex
    Shang, Jingbo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 10456 - 10470