Exploring and Characterizing Large Language Models for Embedded System Development and Debugging

被引:1
|
作者
Englhardt, Zachary [1 ]
Li, Richard [1 ]
Nissanka, Dilini [1 ]
Zhang, Zhihan [1 ]
Narayanswamy, Girish [1 ]
Breda, Joseph [1 ]
Liu, Xin [1 ]
Patel, Shwetak [1 ]
Iyer, Vikram [1 ]
机构
[1] Univ Washington, Seattle, WA 98195 USA
关键词
Large Language Models; GPT; Embedded Systems Development; MAKECODE;
D O I
10.1145/3613905.3650764
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Large language models (LLMs) have shown remarkable abilities to generate code. However, their ability to develop software for physical computing and embedded systems, which requires cross-domain hardware and software knowledge, has not been thoroughly studied. We observe through our experiments and a 15-user pilot study that even when LLMs fail to produce working code, they can generate helpful reasoning about embedded design tasks, as well as specific debugging suggestions for both novice and expert developers. These results highlight the potential to develop AI assistants to dramatically lower the barrier to entry for working with hardware. To evaluate the capabilities and limitations of LLMs, we develop an automated testbench to quantify LLM performance on embedded programming tasks and perform 450 trials. We leverage these findings to analyze how programmers interact with these tools including their productivity and sense of fulfillment and outline a human-AI collaborative workflow for developing and debugging embedded systems.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Extension language automation of embedded system debugging
    Parson D.
    Schlieder B.
    Beatty P.
    Automated Software Engineering, 2002, 9 (01) : 7 - 39
  • [2] The virtual debugging system for embedded software development
    Pan, Y
    Abe, N
    Tanaka, K
    FOURTH INTERNATIONAL CONFERENCE ON VIRTUAL REALITY AND ITS APPLICATIONS IN INDUSTRY, 2004, 5444 : 357 - 364
  • [3] The virtual debugging system for embedded software development
    Pan, Y
    Abe, N
    Tanaka, K
    7TH WORLD MULTICONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL IX, PROCEEDINGS: COMPUTER SCIENCE AND ENGINEERING: II, 2003, : 402 - 407
  • [4] DebugBench: Evaluating Debugging Capability of Large Language Models
    Tian, Runchu
    Ye, Yining
    Qin, Yujia
    Cong, Xin
    Lin, Yankai
    Pan, Yinxu
    Wu, Yesai
    Hui, Haotian
    Liu, Weichuan
    Liu, Zhiyuan
    Sun, Maosong
    Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2024, : 4173 - 4198
  • [5] Automatic building energy model development and debugging using large language models agentic workflow
    Zhang, Liang
    Ford, Vitaly
    Chen, Zhelun
    Chen, Jianli
    ENERGY AND BUILDINGS, 2025, 327
  • [6] Friend or foe? Exploring the implications of large language models on the science system
    Fecher, Benedikt
    Hebing, Marcel
    Laufer, Melissa
    Pohle, Joerg
    Sofsky, Fabian
    AI & SOCIETY, 2023, 40 (2) : 447 - 459
  • [7] Study on Debugging Method Based on Embedded System Development Platform
    Yang Penghao
    Wang Rongliang
    Fan Ziguo
    MANUFACTURING PROCESS AND EQUIPMENT, PTS 1-4, 2013, 694-697 : 2646 - 2650
  • [8] Exploring Large Language Models for Classical Philology
    Riemenschneider, Frederick
    Frank, Anette
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 15181 - 15199
  • [9] Exploring Mathematical Conjecturing with Large Language Models
    Johansson, Moa
    Smallbone, Nicholas
    NEURAL-SYMBOLIC LEARNING AND REASONING 2023, NESY 2023, 2023,
  • [10] Exploring Length Generalization in Large Language Models
    Anil, Cem
    Wu, Yuhuai
    Andreassen, Anders
    Lewkowycz, Aitor
    Misra, Vedant
    Ramasesh, Vinay
    Slone, Ambrose
    Gur-Ari, Guy
    Dyer, Ethan
    Neyshabur, Behnam
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,