RTLLM: An Open-Source Benchmark for Design RTL Generation with Large Language Model

被引:10
|
作者
Lu, Yao [1 ]
Liu, Shang [1 ]
Zhang, Qijun [1 ]
Xie, Zhiyao [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ASP-DAC58780.2024.10473904
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inspired by the recent success of large language models (LLMs) like ChatGPT, researchers start to explore the adoption of LLMs for agile hardware design, such as generating design RTL based on natural-language instructions. However, in existing works, their target designs are all relatively simple and in a small scale, and proposed by the authors themselves, making a fair comparison among different LLM solutions challenging. In addition, many prior works only focus on the design correctness, without evaluating the design qualities of generated design RTL. In this work, we propose an open-source benchmark named RTLLM, for generating design RTL with natural language instructions. To systematically evaluate the auto-generated design RTL, we summarized three progressive goals, named syntax goal, functionality goal, and design quality goal. This benchmark can automatically provide a quantitative evaluation of any given LLM-based solution. Furthermore, we propose an easy-to-use yet surprisingly effective prompt engineering technique named self-planning, which proves to significantly boost the performance of GPT-3.5 in our proposed benchmark.
引用
收藏
页码:722 / 727
页数:6
相关论文
共 50 条
  • [31] The open-source model - Preface
    Birman, A
    Ritsko, JJ
    IBM SYSTEMS JOURNAL, 2005, 44 (02) : 213 - 214
  • [32] PharmaLLM: A Medicine Prescriber Chatbot Exploiting Open-Source Large Language Models
    Ayesha Azam
    Zubaira Naz
    Muhammad Usman Ghani Khan
    Human-Centric Intelligent Systems, 2024, 4 (4): : 527 - 544
  • [33] OpenJustice.ai: A Global Open-Source Legal Language Model
    Dahan, Samuel
    Bhambhoria, Rohan
    Liang, David
    Zhu, Xiaodan
    LEGAL KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 379 : 387 - 390
  • [34] Automated Essay Scoring and Revising Based on Open-Source Large Language Models
    Song, Yishen
    Zhu, Qianta
    Wang, Huaibo
    Zheng, Qinhua
    IEEE TRANSACTIONS ON LEARNING TECHNOLOGIES, 2024, 17 : 1920 - 1930
  • [35] Open-Source Large Language Models in Radiology: A Review and Tutorialfor PracticalResearch and ClinicalDeployment
    Savage, Cody H.
    Kanhere, Adway
    Parekh, Vishwa
    Langlotz, Curtis P.
    Joshi, Anupam
    Huang, Heng
    Doo, Florence X.
    RADIOLOGY, 2025, 314 (01)
  • [36] Open-source large language models in action: A bioinformatics chatbot for PRIDE database
    Bai, Jingwen
    Kamatchinathan, Selvakumar
    Kundu, Deepti J.
    Bandla, Chakradhar
    Vizcaino, Juan Antonio
    Perez-Riverol, Yasset
    PROTEOMICS, 2024, 24 (21-22)
  • [37] Open-source large language models in medical education: Balancing promise and challenges
    Ray, Partha Pratim
    ANATOMICAL SCIENCES EDUCATION, 2024, 17 (06) : 1361 - 1362
  • [38] Benchmarking Open-Source Large Language Models on Code-Switched Tagalog-English Retrieval Augmented Generation
    Adoptante, Aunhel John M.
    Castro, Jasper Adrian Dwight, V
    Medrana, Micholo Lanz B.
    Ocampo, Alyssa Patricia B.
    Peramo, Elmer C.
    Miranda, Melissa Ruth M.
    JOURNAL OF ADVANCES IN INFORMATION TECHNOLOGY, 2025, 16 (02) : 233 - 242
  • [39] Open-source Defect Injection Benchmark Testbed for the Evaluation of Testing
    Bures, Miroslav
    Herout, Pavel
    Ahmed, Bestoun S.
    2020 IEEE 13TH INTERNATIONAL CONFERENCE ON SOFTWARE TESTING, VALIDATION AND VERIFICATION (ICST 2020), 2020, : 442 - 447
  • [40] BulletArm: An Open-Source Robotic Manipulation Benchmark and Learning Framework
    Wang, Dian
    Kohler, Colin
    Zhu, Xupeng
    Jia, Mingxi
    Platt, Robert
    ROBOTICS RESEARCH, ISRR 2022, 2023, 27 : 335 - 350