Discovering Photoswitchable Molecules for Drug Delivery with Large Language Models and Chemist Instruction Training

被引:0
|
作者
Hu, Junjie [1 ]
Wu, Peng [2 ]
Li, Yulin [3 ]
Li, Qi [1 ]
Wang, Shiyi [1 ]
Liu, Yang [4 ]
Qian, Kun [5 ]
Yang, Guang [1 ,6 ,7 ,8 ]
机构
[1] Imperial Coll London, Bioengn Dept & Imperial X, London W12 7SL, England
[2] Ningxia Univ, Sch Chem & Chem Engn, Yinchuan 750014, Peoples R China
[3] Chinese Univ Hong Kong, Dept Math, Shatin, Hong Kong, Peoples R China
[4] Shanxi Med Univ, Shanxi Bethune Hosp, Tongji Shanxi Hosp, Shanxi Acad Med Sci,3 Hosp, Taiyuan 030032, Peoples R China
[5] Fudan Univ, Zhongshan Hosp, Dept Informat & Intelligence Dev, 180 Fenglin Rd, Shanghai 200032, Peoples R China
[6] Imperial Coll London, Natl Heart & Lung Inst, London SW7 2AZ, England
[7] Royal Brompton Hosp, Cardiovasc Res Ctr, London SW3 6NP, England
[8] Kings Coll London, Sch Biomed Engn & Imaging Sci, London WC2R 2LS, England
关键词
drug delivery; photoresponsive molecules; quantum chemistry; language model; RLHF; ABSORPTION; ENERGIES;
D O I
10.3390/ph17101300
中图分类号
R914 [药物化学];
学科分类号
100701 ;
摘要
Background: As large language models continue to expand in size and diversity, their substantial potential and the relevance of their applications are increasingly being acknowledged. The rapid advancement of these models also holds profound implications for the long-term design of stimulus-responsive materials used in drug delivery. Methods: The large model used Hugging Face's Transformers package with BigBird, Gemma, and GPT NeoX architectures. Pre-training used the PubChem dataset, and fine-tuning used QM7b. Chemist instruction training was based on Direct Preference Optimization. Drug Likeness, Synthetic Accessibility, and PageRank Scores were used to filter molecules. All computational chemistry simulations were performed using ORCA and Time-Dependent Density-Functional Theory. Results: To optimize large models for extensive dataset processing and comprehensive learning akin to a chemist's intuition, the integration of deeper chemical insights is imperative. Our study initially compared the performance of BigBird, Gemma, GPT NeoX, and others, specifically focusing on the design of photoresponsive drug delivery molecules. We gathered excitation energy data through computational chemistry tools and further investigated light-driven isomerization reactions as a critical mechanism in drug delivery. Additionally, we explored the effectiveness of incorporating human feedback into reinforcement learning to imbue large models with chemical intuition, enhancing their understanding of relationships involving -N=N- groups in the photoisomerization transitions of photoresponsive molecules. Conclusions: We implemented an efficient design process based on structural knowledge and data, driven by large language model technology, to obtain a candidate dataset of specific photoswitchable molecules. However, the lack of specialized domain datasets remains a challenge for maximizing model performance.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] AI Computing Systems for Large Language Models Training
    Zhang, Zhen-Xing
    Wen, Yuan-Bo
    Lyu, Han-Qi
    Liu, Chang
    Zhang, Rui
    Li, Xia-Qing
    Wang, Chao
    Du, Zi-Dong
    Guo, Qi
    Li, Ling
    Zhou, Xue-Hai
    Chen, Yun-Ji
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2025, 40 (01) : 6 - 41
  • [32] Game of LLMs: Discovering Structural Constructs in Activities using Large Language Models
    Hiremath, Shruthi K.
    Plotz, Thomas
    COMPANION OF THE 2024 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING, UBICOMP COMPANION 2024, 2024, : 487 - 492
  • [33] Discovering sentiment insights: streamlining tourism review analysis with Large Language Models
    Guidotti, Dario
    Pandolfo, Laura
    Pulina, Luca
    INFORMATION TECHNOLOGY & TOURISM, 2025, : 227 - 261
  • [34] Photoswitchable Molecules in Long-Wavelength Light-Responsive Drug Delivery: From Molecular Design to Applications
    Jia, Shiyang
    Fong, Wye-Khay
    Graham, Bim
    Boyd, Ben J.
    CHEMISTRY OF MATERIALS, 2018, 30 (09) : 2873 - 2887
  • [35] Advancing entity recognition in biomedicine via instruction tuning of large language models
    Keloth, Vipina K.
    Hu, Yan
    Xie, Qianqian
    Peng, Xueqing
    Wang, Yan
    Zheng, Andrew
    Selek, Melih
    Raja, Kalpana
    Wei, Chih Hsuan
    Jin, Qiao
    Lu, Zhiyong
    Chen, Qingyu
    Xu, Hua
    BIOINFORMATICS, 2024, 40 (04)
  • [36] Reflective Instruction Tuning: Mitigating Hallucinations in Large Vision-Language Models
    Zhang, Jinrui
    Wang, Teng
    Zhang, Haigang
    Lu, Ping
    Zheng, Feng
    COMPUTER VISION - ECCV 2024, PT XXXVII, 2025, 15095 : 196 - 213
  • [37] Instruction Tuning Large Language Models for Multimodal Relation Extraction Using LoRA
    Li, Zou
    Pang, Ning
    Zhao, Xiang
    WEB INFORMATION SYSTEMS AND APPLICATIONS, WISA 2024, 2024, 14883 : 364 - 376
  • [38] EasyInstruct: An Easy-to-use Instruction Processing Framework for Large Language Models
    Ou, Yixin
    Zhang, Ningyu
    Gui, Honghao
    Xu, Ziwen
    Qiao, Shuofei
    Xue, Yida
    Fang, Runnan
    Liu, Kangwei
    Li, Lei
    Bi, Zhen
    Zheng, Guozhou
    Chen, Huajun
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 3: SYSTEM DEMONSTRATIONS, 2024, : 94 - 106
  • [39] Mitigating Hallucinations in Large Vision-Language Models with Instruction Contrastive Decoding
    Wang, Xintong
    Pan, Jingheng
    Ding, Liang
    Biemann, Chris
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 15840 - 15853
  • [40] WaveCoder: Widespread And Versatile Enhancement For Code Large Language Models By Instruction Tuning
    Yu, Zhaojian
    Zhang, Xin
    Shang, Ning
    Huang, Yangyu
    Xu, Can
    Zhao, Yishujie
    Hu, Wenxiang
    Yin, Qiufeng
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5140 - 5153