DIRECT : A Transformer-based Model for Decompiled Variable Name Recovery

被引:0
|
作者
Nitin, Vikram [1 ]
Saieva, Anthony [1 ]
Ray, Baishakhi [1 ]
Kaiser, Gail [1 ]
机构
[1] Columbia Univ, Dept Comp Sci, New York, NY 10027 USA
来源
NLP4PROG 2021: THE 1ST WORKSHOP ON NATURAL LANGUAGE PROCESSING FOR PROGRAMMING (NLP4PROG 2021) | 2021年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Decompiling binary executables to high-level code is an important step in reverse engineering scenarios, such as malware analysis and legacy code maintenance. However, the generated high-level code is difficult to understand since the original variable names are lost. In this paper, we leverage transformer models to reconstruct the original variable names from decompiled code. Inherent differences between code and natural language present certain challenges in applying conventional transformer-based architectures to variable name recovery. We propose DIRECT, a novel transformer-based architecture customized specifically for the task at hand. We evaluate our model on a dataset of decompiled functions and find that DIRECT outperforms the previous state-of-the-art model by up to 20%. We also present ablation studies evaluating the impact of each of our modifications. We make the source code of DIRECT available to encourage reproducible research.
引用
收藏
页码:48 / 57
页数:10
相关论文
共 50 条
  • [41] Training and analyzing a Transformer-based machine translation model
    Pimentel, Clovis Henrique Martins
    Pires, Thiago Blanch
    TEXTO LIVRE-LINGUAGEM E TECNOLOGIA, 2024, 17
  • [42] A Transformer-Based Fusion Recommendation Model For IPTV Applications
    Li, Heng
    Lei, Hang
    Yang, Maolin
    Zeng, Jinghong
    Zhu, Di
    Fu, Shouwei
    2020 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA (ICAIBD 2020), 2020, : 177 - 182
  • [43] Transformer-based generative adversarial network enabled direct aberration determination
    Chen, Sitong
    Zhu, Zihao
    Wang, Hao
    Wang, Yangyundou
    OPTICAL ENGINEERING, 2024, 63 (06)
  • [44] A Transformer-based Semantic Segmentation Model for Street Fashion Images
    Peng, Dingjie
    Kameyama, Wataru
    INTERNATIONAL WORKSHOP ON ADVANCED IMAGING TECHNOLOGY, IWAIT 2023, 2023, 12592
  • [45] Transformer-Based Molecular Generative Model for Antiviral Drug Design
    Mao, Jiashun
    Wang, Jianmin
    Zeb, Amir
    Cho, Kwang-Hwi
    Jin, Haiyan
    Kim, Jongwan
    Lee, Onju
    Wang, Yunyun
    No, Kyoung Tai
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2023, 64 (07) : 2733 - 2745
  • [46] Transformer-based deep learning model for forced oscillation localization
    Matar, Mustafa
    Estevez, Pablo Gill
    Marchi, Pablo
    Messina, Francisco
    Elmoudi, Ramadan
    Wshah, Safwan
    INTERNATIONAL JOURNAL OF ELECTRICAL POWER & ENERGY SYSTEMS, 2023, 146
  • [47] Transformer-Based Explainable Model for Breast Cancer Lesion Segmentation
    Wang, Huina
    Wei, Lan
    Liu, Bo
    Li, Jianqiang
    Li, Jinshu
    Fang, Juan
    Mooney, Catherine
    APPLIED SCIENCES-BASEL, 2025, 15 (03):
  • [48] Characterization of groundwater contamination: A transformer-based deep learning model
    Bai, Tao
    Tahmasebi, Pejman
    ADVANCES IN WATER RESOURCES, 2022, 164
  • [49] Successful Precipitation Downscaling Through an Innovative Transformer-Based Model
    Yang, Fan
    Ye, Qiaolin
    Wang, Kai
    Sun, Le
    REMOTE SENSING, 2024, 16 (22)
  • [50] A Transformer-based Model for Older Adult Behavior Change Detection
    Akbari, Fateme
    Sartipi, Kamran
    2022 IEEE 10TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2022), 2022, : 27 - 35