Material transformers: deep learning language models for generative materials design

被引:21
|
作者
Fu, Nihang [1 ]
Wei, Lai [1 ]
Song, Yuqi [1 ]
Li, Qinyang [1 ]
Xin, Rui [1 ]
Omee, Sadman Sadeed [1 ]
Dong, Rongzhi [1 ]
Siriwardane, Edirisuriya M. Dilanga [2 ]
Hu, Jianjun [1 ]
机构
[1] Univ South Carolina, Dept Comp Sci & Engn, Columbia, SC 29201 USA
[2] Univ Colombo, Dept Phys, Colombo 03, Sri Lanka
来源
MACHINE LEARNING-SCIENCE AND TECHNOLOGY | 2023年 / 4卷 / 01期
基金
美国国家科学基金会;
关键词
deep learning; language models; generative design; materials discovery; transformer; TOTAL-ENERGY CALCULATIONS; WAVE;
D O I
10.1088/2632-2153/acadcd
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained transformer language models (LMs) on large unlabeled corpus have produced state-of-the-art results in natural language processing, organic molecule design, and protein sequence generation. However, no such models have been applied to learn the composition patterns for the generative design of material compositions. Here we train a series of seven modern transformer models (GPT, GPT-2, GPT-Neo, GPT-J, BLMM, BART, and RoBERTa) for materials design using the expanded formulas of the ICSD, OQMD, and Materials Projects databases. Six different datasets with/out non-charge-neutral or EB samples are used to benchmark the generative design performances and uncover the biases of modern transformer models for the generative design of materials compositions. Our experiments show that the materials transformers based on causal LMs can generate chemically valid material compositions with as high as 97.61% to be charge neutral and 91.22% to be electronegativity balanced, which has more than six times higher enrichment compared to the baseline pseudo-random sampling algorithm. Our LMs also demonstrate high generation novelty and their potential in new materials discovery is proved by their capability to recover the leave-out materials. We also find that the properties of the generated compositions can be tailored by training the models with selected training sets such as high-bandgap samples. Our experiments also show that different models each have their own preference in terms of the properties of the generated samples and their running time complexity varies a lot. We have applied our materials transformers to discover a set of new materials as validated using density functional theory calculations.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Learning Deep Generative Spatial Models for Mobile Robots
    Pronobis, Andrzej
    Rao, Rajesh P. N.
    2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 755 - 762
  • [32] Learning Hierarchical Features from Deep Generative Models
    Zhao, Shengjia
    Song, Jiaming
    Ermon, Stefano
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [33] Semi-supervised Learning with Deep Generative Models
    Kingma, Diederik P.
    Rezende, Danilo J.
    Mohamed, Shakir
    Welling, Max
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [34] Deep learning with the generative models for recommender systems: A survey
    Nahta, Ravi
    Chauhan, Ganpat Singh
    Meena, Yogesh Kumar
    Gopalani, Dinesh
    COMPUTER SCIENCE REVIEW, 2024, 53
  • [35] Flexible and accurate inference and learning for deep generative models
    Vertes, Eszter
    Sahani, Maneesh
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [36] De novo drug design with deep generative models
    Das, Payel
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 2018, 256
  • [37] Application of Deep Generative Models in the Design of Buildings and Structures
    Kanukhin, Aleksandr
    Rynkovskaya, Marina
    PROCEEDINGS OF 6TH INTERNATIONAL CONFERENCE ON CIVIL ENGINEERING AND ARCHITECTURE, VOL 1, ICCEA 2023, 2024, 530 : 58 - 66
  • [38] Author Correction: Physics guided deep learning for generative design of crystal materials with symmetry constraints
    Yong Zhao
    Edirisuriya M. Dilanga Siriwardane
    Zhenyao Wu
    Nihang Fu
    Mohammed Al-Fahdi
    Ming Hu
    Jianjun Hu
    npj Computational Materials, 9
  • [39] Generative Transformers for Design Concept Generation
    Zhu, Qihao
    Luo, Jianxi
    JOURNAL OF COMPUTING AND INFORMATION SCIENCE IN ENGINEERING, 2023, 23 (04)
  • [40] Deep MRI Reconstruction with Generative Vision Transformers
    Korkmaz, Yilmaz
    Yurt, Mahmut
    Dar, Salman Ul Hassan
    Ozbey, Muzaffer
    Cukur, Tolga
    MACHINE LEARNING FOR MEDICAL IMAGE RECONSTRUCTION (MLMIR 2021), 2021, 12964 : 54 - 64