TransPolymer: a Transformer-based language model for polymer property predictions

被引:0
|
作者
Changwen Xu
Yuyang Wang
Amir Barati Farimani
机构
[1] Carnegie Mellon University,Department of Materials Science and Engineering
[2] Carnegie Mellon University,Department of Mechanical Engineering
[3] Carnegie Mellon University,Machine Learning Department
[4] Carnegie Mellon University,Department of Chemical Engineering
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Accurate and efficient prediction of polymer properties is of great significance in polymer design. Conventionally, expensive and time-consuming experiments or simulations are required to evaluate polymer functions. Recently, Transformer models, equipped with self-attention mechanisms, have exhibited superior performance in natural language processing. However, such methods have not been investigated in polymer sciences. Herein, we report TransPolymer, a Transformer-based language model for polymer property prediction. Our proposed polymer tokenizer with chemical awareness enables learning representations from polymer sequences. Rigorous experiments on ten polymer property prediction benchmarks demonstrate the superior performance of TransPolymer. Moreover, we show that TransPolymer benefits from pretraining on large unlabeled dataset via Masked Language Modeling. Experimental results further manifest the important role of self-attention in modeling polymer sequences. We highlight this model as a promising computational tool for promoting rational polymer design and understanding structure-property relationships from a data science view.
引用
收藏
相关论文
共 50 条
  • [1] TransPolymer: a Transformer-based language model for polymer property predictions
    Xu, Changwen
    Wang, Yuyang
    Farimani, Amir Barati
    NPJ COMPUTATIONAL MATERIALS, 2023, 9 (01)
  • [2] High entropy alloy property predictions using a transformer-based language model
    Spyros Kamnis
    Konstantinos Delibasis
    Scientific Reports, 15 (1)
  • [3] Transformer-Based Mechanical Property Prediction for Polymer Matrix Composites
    Lee, Jaewook
    Son, Jinkyung
    Lim, Juri
    Kim, In
    Kim, Seonwoo
    Cho, Namjung
    Choi, Woojin
    Shin, Dongil
    KOREAN JOURNAL OF CHEMICAL ENGINEERING, 2024, 41 (11) : 3005 - 3018
  • [4] Transformer-based heart language model with electrocardiogram annotations
    Tudjarski, Stojancho
    Gusev, Marjan
    Kanoulas, Evangelos
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [5] LVBERT: Transformer-Based Model for Latvian Language Understanding
    Znotins, Arturs
    Barzdins, Guntis
    HUMAN LANGUAGE TECHNOLOGIES - THE BALTIC PERSPECTIVE (HLT 2020), 2020, 328 : 111 - 115
  • [6] ParsBERT: Transformer-based Model for Persian Language Understanding
    Mehrdad Farahani
    Mohammad Gharachorloo
    Marzieh Farahani
    Mohammad Manthouri
    Neural Processing Letters, 2021, 53 : 3831 - 3847
  • [7] ParsBERT: Transformer-based Model for Persian Language Understanding
    Farahani, Mehrdad
    Gharachorloo, Mohammad
    Farahani, Marzieh
    Manthouri, Mohammad
    NEURAL PROCESSING LETTERS, 2021, 53 (06) : 3831 - 3847
  • [8] AN EMPIRICAL STUDY OF TRANSFORMER-BASED NEURAL LANGUAGE MODEL ADAPTATION
    Li, Ke
    Liu, Zhe
    He, Tianxing
    Huang, Hongzhao
    Peng, Fuchun
    Povey, Daniel
    Khudanpur, Sanjeev
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7934 - 7938
  • [9] Transformer-Based Single-Cell Language Model: A Survey
    Lan, Wei
    He, Guohang
    Liu, Mingyang
    Chen, Qingfeng
    Cao, Junyue
    Peng, Wei
    BIG DATA MINING AND ANALYTICS, 2024, 7 (04): : 1169 - 1186
  • [10] Generating Qualitative Descriptions of Diagrams with a Transformer-Based Language Model
    Schorlemmer, Marco
    Ballout, Mohamad
    Kuehnberger, Kai-Uwe
    DIAGRAMMATIC REPRESENTATION AND INFERENCE, DIAGRAMS 2024, 2024, 14981 : 61 - 75