Material transformers: deep learning language models for generative materials design

被引:21
|
作者
Fu, Nihang [1 ]
Wei, Lai [1 ]
Song, Yuqi [1 ]
Li, Qinyang [1 ]
Xin, Rui [1 ]
Omee, Sadman Sadeed [1 ]
Dong, Rongzhi [1 ]
Siriwardane, Edirisuriya M. Dilanga [2 ]
Hu, Jianjun [1 ]
机构
[1] Univ South Carolina, Dept Comp Sci & Engn, Columbia, SC 29201 USA
[2] Univ Colombo, Dept Phys, Colombo 03, Sri Lanka
来源
MACHINE LEARNING-SCIENCE AND TECHNOLOGY | 2023年 / 4卷 / 01期
基金
美国国家科学基金会;
关键词
deep learning; language models; generative design; materials discovery; transformer; TOTAL-ENERGY CALCULATIONS; WAVE;
D O I
10.1088/2632-2153/acadcd
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained transformer language models (LMs) on large unlabeled corpus have produced state-of-the-art results in natural language processing, organic molecule design, and protein sequence generation. However, no such models have been applied to learn the composition patterns for the generative design of material compositions. Here we train a series of seven modern transformer models (GPT, GPT-2, GPT-Neo, GPT-J, BLMM, BART, and RoBERTa) for materials design using the expanded formulas of the ICSD, OQMD, and Materials Projects databases. Six different datasets with/out non-charge-neutral or EB samples are used to benchmark the generative design performances and uncover the biases of modern transformer models for the generative design of materials compositions. Our experiments show that the materials transformers based on causal LMs can generate chemically valid material compositions with as high as 97.61% to be charge neutral and 91.22% to be electronegativity balanced, which has more than six times higher enrichment compared to the baseline pseudo-random sampling algorithm. Our LMs also demonstrate high generation novelty and their potential in new materials discovery is proved by their capability to recover the leave-out materials. We also find that the properties of the generated compositions can be tailored by training the models with selected training sets such as high-bandgap samples. Our experiments also show that different models each have their own preference in terms of the properties of the generated samples and their running time complexity varies a lot. We have applied our materials transformers to discover a set of new materials as validated using density functional theory calculations.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] Deep Generative Models in Engineering Design: A Review
    Regenwetter, Lyle
    Nobari, Amin Heyrani
    Ahmed, Faez
    JOURNAL OF MECHANICAL DESIGN, 2022, 144 (07)
  • [22] Generative Deep Neural Networks for Inverse Materials Design Using Backpropagation and Active Learning
    Chen, Chun-Teh
    Gu, Grace X.
    ADVANCED SCIENCE, 2020, 7 (05)
  • [23] Generative design of stable semiconductor materials using deep learning and density functional theory
    Siriwardane, Edirisuriya M. Dilanga
    Zhao, Yong
    Perera, Indika
    Hu, Jianjun
    NPJ COMPUTATIONAL MATERIALS, 2022, 8 (01)
  • [24] Generative design of stable semiconductor materials using deep learning and density functional theory
    Edirisuriya M. Dilanga Siriwardane
    Yong Zhao
    Indika Perera
    Jianjun Hu
    npj Computational Materials, 8
  • [25] Generative Deep Learning for Targeted Compound Design
    Sousa, Tiago
    Correia, Joao
    Pereira, Vitor
    Rocha, Miguel
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2021, 61 (11) : 5343 - 5361
  • [26] DzAIN: Deep learning based generative design
    Kallioras, Nikos Ath.
    Lagaros, Nikos D.
    1ST INTERNATIONAL CONFERENCE ON OPTIMIZATION-DRIVEN ARCHITECTURAL DESIGN (OPTARCH 2019), 2020, 44 : 591 - 598
  • [27] Deep-learning generative models enable design of synthetic orthologs of a signaling protein
    Lian, Xinran
    Praljak, Niksa
    Ferguson, Andrew L.
    Ranganathan, Rama
    BIOPHYSICAL JOURNAL, 2023, 122 (03) : 311A - 311A
  • [28] Computational Discovery of New 2D Materials Using Deep Learning Generative Models
    Song, Yuqi
    Siriwardane, Edirisuriya M. Dilanga
    Zhao, Yong
    Hu, Jianjun
    ACS APPLIED MATERIALS & INTERFACES, 2021, 13 (45) : 53303 - 53313
  • [29] Unsupervised learning of global factors in deep generative models
    Peis, Ignacio
    Olmos, Pablo M.
    Artes-Rodriguez, Antonio
    PATTERN RECOGNITION, 2022, 134
  • [30] Wasserstein Learning of Deep Generative Point Process Models
    Xiao, Shuai
    Farajtabar, Mehrdad
    Ye, Xiaojing
    Yang, Junchi
    Song, Le
    Zha, Hongyuan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30