SemiH: DFT Hamiltonian neural network training with semi-supervised learning

被引:0
|
作者
Cho, Yucheol [1 ]
Choi, Guenseok [1 ]
Ham, Gyeongdo [1 ]
Shin, Mincheol [1 ]
Kim, Daeshik [1 ]
机构
[1] Korea Adv Inst Sci & Technol KAIST, Sch Elect Engn, Daejeon 34141, South Korea
来源
关键词
density functional theory; neural network Hamiltonian; message-passing neural network; semi-supervised learning; unlabeled data; pseudo Hamiltonian; graph neural network; DENSITY-FUNCTIONAL THEORY; ELECTRONIC-STRUCTURE;
D O I
10.1088/2632-2153/ad7227
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Over the past decades, density functional theory (DFT) calculations have been utilized in various fields such as materials science and semiconductor devices. However, due to the inherent nature of DFT calculations, which rigorously consider interactions between atoms, they require significant computational cost. To address this, extensive research has recently focused on training neural networks to replace DFT calculations. However, previous methods for training neural networks necessitated an extensive number of DFT simulations to acquire the ground truth (Hamiltonians). Conversely, when dealing with a limited amount of training data, deep learning models often display increased errors in predicting Hamiltonians and band structures for testing data. This phenomenon poses the potential risk of generating inaccurate physical interpretations, including the emergence of unphysical branches within band structures. To tackle this challenge, we propose a novel deep learning-based method for calculating DFT Hamiltonians, specifically tailored to produce accurate results with limited training data. Our framework not only employs supervised learning with the calculated Hamiltonian but also generates pseudo Hamiltonians (targets for unlabeled data) and trains the neural networks on unlabeled data. Particularly, our approach, which leverages unlabeled data, is noteworthy as it marks the first attempt in the field of neural network Hamiltonians. Our framework showcases the superior performance of our framework compared to the state-of-the-art approach across various datasets, such as MoS2, Bi2Te3, HfO2, and InGaAs. Moreover, our framework demonstrates enhanced generalization performance by effectively utilizing unlabeled data, achieving noteworthy results when evaluated on data more complex than the training set, such as configurations with more atoms and temperature ranges outside the training data.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] A Deep Neural Network Based on ELM for Semi-supervised Learning of Image Classification
    Peiju Chang
    Jiangshe Zhang
    Junying Hu
    Zengjie Song
    Neural Processing Letters, 2018, 48 : 375 - 388
  • [32] Semi-supervised node classification via graph learning convolutional neural network
    Li, Kangjie
    Ye, Wenjing
    APPLIED INTELLIGENCE, 2022, 52 (11) : 12724 - 12736
  • [33] Gated Relational Graph Neural Network for Semi-supervised Learning on Knowledge Graphs
    Chen, Yuyan
    Zou, Lei
    Qin, Zongyue
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2019, 2019, 11881 : 617 - 629
  • [34] A Deep Neural Network Based on ELM for Semi-supervised Learning of Image Classification
    Chang, Peiju
    Zhang, Jiangshe
    Hu, Junying
    Song, Zengjie
    NEURAL PROCESSING LETTERS, 2018, 48 (01) : 375 - 388
  • [35] Semi-supervised neural network training method for fast-moving object detection
    Sevo, Igor
    2018 14TH SYMPOSIUM ON NEURAL NETWORKS AND APPLICATIONS (NEUREL), 2018,
  • [36] Semi-supervised Maximum Mutual Information Training of Deep Neural Network Acoustic Models
    Manohar, Vimal
    Povey, Daniel
    Khudanpur, Sanjeev
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 2630 - 2634
  • [37] DEEP NEURAL NETWORK FEATURES AND SEMI-SUPERVISED TRAINING FOR LOW RESOURCE SPEECH RECOGNITION
    Thomas, Samuel
    Seltzer, Michael L.
    Church, Kenneth
    Hermansky, Hynek
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 6704 - 6708
  • [38] DMT: Dynamic mutual training for semi-supervised learning
    Feng, Zhengyang
    Zhou, Qianyu
    Gu, Qiqi
    Tan, Xin
    Cheng, Guangliang
    Lu, Xuequan
    Shi, Jianping
    Ma, Lizhuang
    PATTERN RECOGNITION, 2022, 130
  • [39] DMT: Dynamic mutual training for semi-supervised learning
    Feng, Zhengyang
    Zhou, Qianyu
    Gu, Qiqi
    Tan, Xin
    Cheng, Guangliang
    Lu, Xuequan
    Shi, Jianping
    Ma, Lizhuang
    PATTERN RECOGNITION, 2022, 130
  • [40] GraphMix: Improved Training of GNNs for Semi-Supervised Learning
    Verma, Vikas
    Qu, Meng
    Kawaguchi, Kenji
    Lamb, Alex
    Bengio, Yoshua
    Kannala, Juho
    Tang, Jian
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10024 - 10032