Discovering Photoswitchable Molecules for Drug Delivery with Large Language Models and Chemist Instruction Training

被引:0
|
作者
Hu, Junjie [1 ]
Wu, Peng [2 ]
Li, Yulin [3 ]
Li, Qi [1 ]
Wang, Shiyi [1 ]
Liu, Yang [4 ]
Qian, Kun [5 ]
Yang, Guang [1 ,6 ,7 ,8 ]
机构
[1] Imperial Coll London, Bioengn Dept & Imperial X, London W12 7SL, England
[2] Ningxia Univ, Sch Chem & Chem Engn, Yinchuan 750014, Peoples R China
[3] Chinese Univ Hong Kong, Dept Math, Shatin, Hong Kong, Peoples R China
[4] Shanxi Med Univ, Shanxi Bethune Hosp, Tongji Shanxi Hosp, Shanxi Acad Med Sci,3 Hosp, Taiyuan 030032, Peoples R China
[5] Fudan Univ, Zhongshan Hosp, Dept Informat & Intelligence Dev, 180 Fenglin Rd, Shanghai 200032, Peoples R China
[6] Imperial Coll London, Natl Heart & Lung Inst, London SW7 2AZ, England
[7] Royal Brompton Hosp, Cardiovasc Res Ctr, London SW3 6NP, England
[8] Kings Coll London, Sch Biomed Engn & Imaging Sci, London WC2R 2LS, England
关键词
drug delivery; photoresponsive molecules; quantum chemistry; language model; RLHF; ABSORPTION; ENERGIES;
D O I
10.3390/ph17101300
中图分类号
R914 [药物化学];
学科分类号
100701 ;
摘要
Background: As large language models continue to expand in size and diversity, their substantial potential and the relevance of their applications are increasingly being acknowledged. The rapid advancement of these models also holds profound implications for the long-term design of stimulus-responsive materials used in drug delivery. Methods: The large model used Hugging Face's Transformers package with BigBird, Gemma, and GPT NeoX architectures. Pre-training used the PubChem dataset, and fine-tuning used QM7b. Chemist instruction training was based on Direct Preference Optimization. Drug Likeness, Synthetic Accessibility, and PageRank Scores were used to filter molecules. All computational chemistry simulations were performed using ORCA and Time-Dependent Density-Functional Theory. Results: To optimize large models for extensive dataset processing and comprehensive learning akin to a chemist's intuition, the integration of deeper chemical insights is imperative. Our study initially compared the performance of BigBird, Gemma, GPT NeoX, and others, specifically focusing on the design of photoresponsive drug delivery molecules. We gathered excitation energy data through computational chemistry tools and further investigated light-driven isomerization reactions as a critical mechanism in drug delivery. Additionally, we explored the effectiveness of incorporating human feedback into reinforcement learning to imbue large models with chemical intuition, enhancing their understanding of relationships involving -N=N- groups in the photoisomerization transitions of photoresponsive molecules. Conclusions: We implemented an efficient design process based on structural knowledge and data, driven by large language model technology, to obtain a candidate dataset of specific photoswitchable molecules. However, the lack of specialized domain datasets remains a challenge for maximizing model performance.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Discovering Research Areas in Dataset Applications through Knowledge Graphs and Large Language Models
    Gerasimov, Irina
    Mehrabian, Armin
    Kc, Binita
    Alfred, Jerome
    Mcguire, Michael P.
    Proceedings - 2024 IEEE 20th International Conference on e-Science, e-Science 2024, 2024,
  • [42] Discovering Research Areas in Dataset Applications through Knowledge Graphs and Large Language Models
    Gerasimov, Irina
    Mehrabian, Armin
    Binita, K. C.
    Alfred, Jerome
    McGuire, Michael P.
    2024 IEEE 20TH INTERNATIONAL CONFERENCE ON E-SCIENCE, E-SCIENCE 2024, 2024,
  • [43] A Survey of Metrics to Enhance Training Dependability in Large Language Models
    Fang, Wenyi
    Zhang, Hao
    Gong, Ziyu
    Zeng, Longbin
    Lu, Xuhui
    Liu, Biao
    Wu, Xiaoyu
    Zheng, Yang
    Hu, Zheng
    Zhang, Xun
    2023 IEEE 34TH INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING WORKSHOPS, ISSREW, 2023, : 180 - 185
  • [44] Large language models reshaping molecular biology and drug development
    Tripathi, Satvik
    Gabriel, Kyla
    Tripathi, Pushpendra Kumar
    Kim, Edward
    CHEMICAL BIOLOGY & DRUG DESIGN, 2024, 103 (06)
  • [45] Unlocking the Capabilities of Large Language Models for Accelerating Drug Development
    Anderson, Wes
    Braun, Ian
    Bhatnagar, Roopal
    Romero, Klaus
    Walls, Ramona
    Schito, Marco
    Podichetty, Jagdeep T.
    CLINICAL PHARMACOLOGY & THERAPEUTICS, 2024, 116 (01) : 38 - 41
  • [46] Integrating large language models and generative artificial intelligence tools into information literacy instruction
    Carroll, Alexander J.
    Borycz, Joshua
    JOURNAL OF ACADEMIC LIBRARIANSHIP, 2024, 50 (04):
  • [47] Toward the Evaluation of Large Language Models Considering Score Variance across Instruction Templates
    Sakai, Yusuke
    Nohejl, Adam
    Hang, Jiangnan
    Kamigaito, Hidetaka
    Watanabe, Taro
    arXiv,
  • [48] Cendol: Open Instruction-tuned Generative Large Language Models for Indonesian Languages
    Cahyawijaya, Samuel
    Lovenia, Holy
    Koto, Fajri
    Pu, Rifki Afina
    Dave, Emmanuel
    Leeo, Jhonson
    Shadiee, Nuur
    Cenggore, Wawan
    Akbar, Salsabil Maulana
    Mahendre, Muhammad Ihza
    Putr, Dea Annisayanti
    Wile, Bryan
    Winata, Genta Indra
    Om, Alham Fikri
    Purwarianti, Ayu
    Fungi, Pascale
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 14899 - 14914
  • [49] FUZZLLM: A NOVEL AND UNIVERSAL FUZZING FRAMEWORK FOR PROACTIVELY DISCOVERING JAILBREAK VULNERABILITIES IN LARGE LANGUAGE MODELS
    Yao, Dongyu
    Zhang, Jianshu
    Harris, Ian G.
    Carlsson, Marcel
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 4485 - 4489
  • [50] Predictive Prompts with Joint Training of Large Language Models for Explainable Recommendation
    Lin, Ching-Sheng
    Tsai, Chung-Nan
    Su, Shao-Tang
    Jwo, Jung-Sing
    Lee, Cheng-Hsiung
    Wang, Xin
    MATHEMATICS, 2023, 11 (20)