DolphCoder: Echo-Locating Code Large Language Models with Diverse and Multi-Objective Instruction Tuning

被引:0
|
作者
Wang, Yejie [1 ]
He, Keqing [2 ]
Dong, Guanting [1 ]
Wang, Pei [1 ]
Zeng, Weihao [1 ]
Diao, Muxi [1 ]
Zhang, Mengdi [2 ]
Wang, Jingang [2 ]
Cai, Xunliang [2 ]
Xu, Weiran [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
[2] Meituan, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Code Large Language Models (Code LLMs) have demonstrated outstanding performance in code-related tasks. Several instruction tuning approaches have been proposed to boost the code generation performance of pre-trained Code LLMs. In this paper, we introduce a diverse instruction model (DolphCoder) with self-evaluating for code generation. It learns diverse instruction targets and combines a code evaluation objective to enhance its code generation ability. Our model achieves superior performance on the HumanEval and MBPP benchmarks, demonstrating new insights for future code instruction tuning work. Our key findings are: (1) Augmenting more diverse responses with distinct reasoning paths increases the code capability of LLMs. (2) Improving one's ability to evaluate the correctness of code solutions also enhances their ability to create it. Our code is released at https://github.com/prisnlp/DolphCoder.
引用
收藏
页码:4706 / 4721
页数:16
相关论文
共 50 条
  • [31] Fine-tuning and prompt engineering for large language models-based code review automation
    Pornprasit, Chanathip
    Tantithamthavorn, Chakkrit
    INFORMATION AND SOFTWARE TECHNOLOGY, 2024, 175
  • [32] Double-Fine-Tuning Multi-Objective Vision-and-Language Transformer for Social Media Popularity Prediction
    Chen, Xiaolu
    Chen, Weilong
    Huang, Chenghao
    Zhang, Zhongjian
    Duan, Lixin
    Zhang, Yanru
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 9462 - 9466
  • [33] EcomGPT: Instruction-Tuning Large Language Models with Chain-of-Task Tasks for E-commerce
    Li, Yangning
    Ma, Shirong
    Wang, Xiaobin
    Huang, Shen
    Jiang, Chengyue
    Zheng, Hai-Tao
    Xie, Pengjun
    Huang, Fei
    Jiang, Yong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 18582 - 18590
  • [34] Multi-turn Instruction Invocation on Human-Robot Interaction by Large Language Models
    Cheng, Baoping
    Huang, Yong
    Sun, Xiaoran
    Hu, Jingxi
    Li, Bo
    Pu, Qiran
    Wu, Zijian
    Tao, Xiaoming
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2024, PT VII, 2025, 15207 : 207 - 219
  • [35] Design of Information Granulation-Based Fuzzy Models with the Aid of Multi-objective Optimization and Successive Tuning Method
    Huang, Wei
    Oh, Sung-Kwun
    Kim, Jeong-Tae
    ADVANCES IN NEURAL NETWORKS - ISNN 2011, PT III, 2011, 6677 : 256 - +
  • [36] Multi-objective reverse engineering of variability-safe feature models based on code dependencies of system variants
    Wesley K. G. Assunção
    Roberto E. Lopez-Herrejon
    Lukas Linsbauer
    Silvia R. Vergilio
    Alexander Egyed
    Empirical Software Engineering, 2017, 22 : 1763 - 1794
  • [37] Multi-objective reverse engineering of variability-safe feature models based on code dependencies of system variants
    Assuncao, Wesley K. G.
    Lopez-Herrejon, Roberto E.
    Linsbauer, Lukas
    Vergilio, Silvia R.
    Egyed, Alexander
    EMPIRICAL SOFTWARE ENGINEERING, 2017, 22 (04) : 1763 - 1794
  • [38] Multi-modal multi-objective model-based genetic programming to find multiple diverse high-quality models
    Sijben, E. M. C.
    Alderliesten, T.
    Bosman, P. A. N.
    PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE (GECCO'22), 2022, : 440 - 448
  • [39] Understanding Multi -Dimensional Efficiency of Fine -Tuning Large Language Models Using SpeedUp, MemoryUp, and EnergyUp
    Chen, Dayuan
    Soto, Noe
    Tuttle, Jonas F.
    Zong, Ziliang
    2024 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS, IPDPSW 2024, 2024, : 929 - 937
  • [40] Multi-objective math problem generation using large language model through an adaptive multi-level retrieval augmentation framework
    Sun, Jianwen
    Shi, Wangzi
    Shen, Xiaoxuan
    Liu, Shengyingjie
    Wei, Luona
    Wan, Qian
    INFORMATION FUSION, 2025, 119