Material transformers: deep learning language models for generative materials design

被引:21
|
作者
Fu, Nihang [1 ]
Wei, Lai [1 ]
Song, Yuqi [1 ]
Li, Qinyang [1 ]
Xin, Rui [1 ]
Omee, Sadman Sadeed [1 ]
Dong, Rongzhi [1 ]
Siriwardane, Edirisuriya M. Dilanga [2 ]
Hu, Jianjun [1 ]
机构
[1] Univ South Carolina, Dept Comp Sci & Engn, Columbia, SC 29201 USA
[2] Univ Colombo, Dept Phys, Colombo 03, Sri Lanka
来源
MACHINE LEARNING-SCIENCE AND TECHNOLOGY | 2023年 / 4卷 / 01期
基金
美国国家科学基金会;
关键词
deep learning; language models; generative design; materials discovery; transformer; TOTAL-ENERGY CALCULATIONS; WAVE;
D O I
10.1088/2632-2153/acadcd
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained transformer language models (LMs) on large unlabeled corpus have produced state-of-the-art results in natural language processing, organic molecule design, and protein sequence generation. However, no such models have been applied to learn the composition patterns for the generative design of material compositions. Here we train a series of seven modern transformer models (GPT, GPT-2, GPT-Neo, GPT-J, BLMM, BART, and RoBERTa) for materials design using the expanded formulas of the ICSD, OQMD, and Materials Projects databases. Six different datasets with/out non-charge-neutral or EB samples are used to benchmark the generative design performances and uncover the biases of modern transformer models for the generative design of materials compositions. Our experiments show that the materials transformers based on causal LMs can generate chemically valid material compositions with as high as 97.61% to be charge neutral and 91.22% to be electronegativity balanced, which has more than six times higher enrichment compared to the baseline pseudo-random sampling algorithm. Our LMs also demonstrate high generation novelty and their potential in new materials discovery is proved by their capability to recover the leave-out materials. We also find that the properties of the generated compositions can be tailored by training the models with selected training sets such as high-bandgap samples. Our experiments also show that different models each have their own preference in terms of the properties of the generated samples and their running time complexity varies a lot. We have applied our materials transformers to discover a set of new materials as validated using density functional theory calculations.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Language Modeling with Deep Transformers
    Irie, Kazuki
    Zeyer, Albert
    Schlueter, Ralf
    Ney, Hermann
    INTERSPEECH 2019, 2019, : 3905 - 3909
  • [42] Generative Deep Learning for Visual Animation in Landscapes Design
    Ardhianto, Peter
    Santosa, Yonathan Purbo
    Moniaga, Christian
    Utami, Maya Putri
    Dewi, Christine
    Christanto, Henoch Juli
    Chen, Abbott Po Shun
    Scientific Programming, 2023, 2023
  • [43] Composition Based Oxidation State Prediction of Materials Using Deep Learning Language Models
    Fu, Nihang
    Hu, Jeffrey
    Feng, Ying
    Morrison, Gregory
    Loye, Hans-Conrad Zur
    Hu, Jianjun
    ADVANCED SCIENCE, 2023, 10 (28)
  • [44] Crystal Composition Transformer: Self-Learning Neural Language Model for Generative and Tinkering Design of Materials
    Wei, Lai
    Li, Qinyang
    Song, Yuqi
    Stefanov, Stanislav
    Dong, Rongzhi
    Fu, Nihang
    Siriwardane, Edirisuriya M. D.
    Chen, Fanglin
    Hu, Jianjun
    ADVANCED SCIENCE, 2024, 11 (36)
  • [45] Reflection on the use of Generative Language Models as a tool for teaching design
    do Amaral, Ines
    VIII IEEE WORLD ENGINEERING EDUCATION CONFERENCE, EDUNINE 2024, 2024,
  • [46] Transformers and genome language models
    Consens, Micaela E.
    Dufault, Cameron
    Wainberg, Michael
    Forster, Duncan
    Karimzadeh, Mehran
    Goodarzi, Hani
    Theis, Fabian J.
    Moses, Alan
    Wang, Bo
    NATURE MACHINE INTELLIGENCE, 2025, : 346 - 362
  • [47] Deep Generative Learning Models for Cloud Intrusion Detection Systems
    Ly Vu
    Quang Uy Nguyen
    Nguyen, N. Diep
    Dinh Thai Hoang
    Dutkiewicz, Eryk
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (01) : 565 - 577
  • [48] Turbulence scaling from deep learning diffusion generative models
    Whittaker, Tim
    Janik, Romuald A.
    Oz, Yaron
    JOURNAL OF COMPUTATIONAL PHYSICS, 2024, 514
  • [49] Semi-Supervised Learning for Deep Causal Generative Models
    Ibrahim, Yasin
    Warr, Hermione
    Kamnitsas, Konstantinos
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT XII, 2024, 15012 : 294 - 303
  • [50] Discovering Binary Codes for Documents by Learning Deep Generative Models
    Hinton, Geoffrey
    Salakhutdinov, Ruslan
    TOPICS IN COGNITIVE SCIENCE, 2011, 3 (01) : 74 - 91