共 50 条
- [2] Evaluating the Language Abilities of Large Language Models vs. Humans: Three Caveats BIOLINGUISTICS, 2024, 18
- [3] Invited Paper: VerilogEval: Evaluating Large Language Models for Verilog Code Generation 2023 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2023,
- [4] JavaBench: A Benchmark of Object-Oriented Code Generation for Evaluating Large Language Models Proceedings - 2024 39th ACM/IEEE International Conference on Automated Software Engineering, ASE 2024, : 870 - 882
- [5] VHDL-Eval: A Framework for Evaluating Large Language Models in VHDL Code Generation 2024 IEEE LLM AIDED DESIGN WORKSHOP, LAD 2024, 2024,
- [8] Quantifying Contamination in Evaluating Code Generation Capabilities of Language Models PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 14116 - 14137
- [9] Evaluating Large Language Models on Controlled Generation Tasks 2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3155 - 3168
- [10] Natural Language Dataset Generation Framework for Visualizations Powered by Large Language Models PROCEEDINGS OF THE 2024 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYTEMS (CHI 2024), 2024,