SemiH: DFT Hamiltonian neural network training with semi-supervised learning

被引:0
|
作者
Cho, Yucheol [1 ]
Choi, Guenseok [1 ]
Ham, Gyeongdo [1 ]
Shin, Mincheol [1 ]
Kim, Daeshik [1 ]
机构
[1] Korea Adv Inst Sci & Technol KAIST, Sch Elect Engn, Daejeon 34141, South Korea
来源
关键词
density functional theory; neural network Hamiltonian; message-passing neural network; semi-supervised learning; unlabeled data; pseudo Hamiltonian; graph neural network; DENSITY-FUNCTIONAL THEORY; ELECTRONIC-STRUCTURE;
D O I
10.1088/2632-2153/ad7227
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Over the past decades, density functional theory (DFT) calculations have been utilized in various fields such as materials science and semiconductor devices. However, due to the inherent nature of DFT calculations, which rigorously consider interactions between atoms, they require significant computational cost. To address this, extensive research has recently focused on training neural networks to replace DFT calculations. However, previous methods for training neural networks necessitated an extensive number of DFT simulations to acquire the ground truth (Hamiltonians). Conversely, when dealing with a limited amount of training data, deep learning models often display increased errors in predicting Hamiltonians and band structures for testing data. This phenomenon poses the potential risk of generating inaccurate physical interpretations, including the emergence of unphysical branches within band structures. To tackle this challenge, we propose a novel deep learning-based method for calculating DFT Hamiltonians, specifically tailored to produce accurate results with limited training data. Our framework not only employs supervised learning with the calculated Hamiltonian but also generates pseudo Hamiltonians (targets for unlabeled data) and trains the neural networks on unlabeled data. Particularly, our approach, which leverages unlabeled data, is noteworthy as it marks the first attempt in the field of neural network Hamiltonians. Our framework showcases the superior performance of our framework compared to the state-of-the-art approach across various datasets, such as MoS2, Bi2Te3, HfO2, and InGaAs. Moreover, our framework demonstrates enhanced generalization performance by effectively utilizing unlabeled data, achieving noteworthy results when evaluated on data more complex than the training set, such as configurations with more atoms and temperature ranges outside the training data.
引用
收藏
页数:16
相关论文
共 50 条
  • [11] Manifold adversarial training for supervised and semi-supervised learning
    Zhang, Shufei
    Huang, Kaizhu
    Zhu, Jianke
    Liu, Yang
    NEURAL NETWORKS, 2021, 140 : 282 - 293
  • [12] SEMI-SUPERVISED TRAINING OF DEEP NEURAL NETWORKS
    Vesely, Karel
    Hannemann, Mirko
    Burget, Lukas
    2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2013, : 267 - 272
  • [13] Self-Training using Selection Network for Semi-supervised Learning
    Jeong, Jisoo
    Lee, Seungeui
    Kwak, Nojun
    ICPRAM: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION APPLICATIONS AND METHODS, 2020, : 23 - 32
  • [14] Semi-Supervised Learning with the Integration of Fuzzy Clustering and Artificial Neural Network
    Saha, Indrajit
    Debnath, Nivriti
    HYBRID INTELLIGENT SYSTEMS, HIS 2017, 2018, 734 : 20 - 29
  • [15] A Reflex Fuzzy Min Max Neural Network for Semi-supervised Learning
    Nandedkar, A.
    Biswas, P.
    JOURNAL OF INTELLIGENT SYSTEMS, 2008, 17 (1-3) : 5 - 17
  • [16] COSNet: A Cost Sensitive Neural Network for Semi-supervised Learning in Graphs
    Bertoni, Alberto
    Frasca, Marco
    Valentini, Giorgio
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT I, 2011, 6911 : 219 - 234
  • [17] On the Learning Dynamics of Semi-Supervised Training for ASR
    Wallington, Electra
    Kershenbaum, Benji
    Klejch, Ondrej
    Bell, Peter
    INTERSPEECH 2021, 2021, : 716 - 720
  • [18] Interpolation Consistency Training for Semi-Supervised Learning
    Verma, Vikas
    Lamb, Alex
    Kannala, Juho
    Bengio, Yoshua
    Lopez-Paz, David
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3635 - 3641
  • [19] Interpolation consistency training for semi-supervised learning
    Verma, Vikas
    Kawaguchi, Kenji
    Lamb, Alex
    Kannala, Juho
    Solin, Arno
    Bengio, Yoshua
    Lopez-Paz, David
    NEURAL NETWORKS, 2022, 145 : 90 - 106
  • [20] MarginGAN: Adversarial Training in Semi-Supervised Learning
    Dong, Jinhao
    Lin, Tong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32