DolphCoder: Echo-Locating Code Large Language Models with Diverse and Multi-Objective Instruction Tuning

被引:0
|
作者
Wang, Yejie [1 ]
He, Keqing [2 ]
Dong, Guanting [1 ]
Wang, Pei [1 ]
Zeng, Weihao [1 ]
Diao, Muxi [1 ]
Zhang, Mengdi [2 ]
Wang, Jingang [2 ]
Cai, Xunliang [2 ]
Xu, Weiran [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
[2] Meituan, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Code Large Language Models (Code LLMs) have demonstrated outstanding performance in code-related tasks. Several instruction tuning approaches have been proposed to boost the code generation performance of pre-trained Code LLMs. In this paper, we introduce a diverse instruction model (DolphCoder) with self-evaluating for code generation. It learns diverse instruction targets and combines a code evaluation objective to enhance its code generation ability. Our model achieves superior performance on the HumanEval and MBPP benchmarks, demonstrating new insights for future code instruction tuning work. Our key findings are: (1) Augmenting more diverse responses with distinct reasoning paths increases the code capability of LLMs. (2) Improving one's ability to evaluate the correctness of code solutions also enhances their ability to create it. Our code is released at https://github.com/prisnlp/DolphCoder.
引用
收藏
页码:4706 / 4721
页数:16
相关论文
共 50 条
  • [1] OCTOPACK: INSTRUCTION TUNING CODE LARGE LANGUAGE MODELS
    Muennighoff, Niklas
    Liu, Qian
    Zebaze, Armel
    Zheng, Qinkai
    Hui, Binyuan
    Zhuo, Terry Yue
    Singh, Swayam
    Tang, Xiangru
    von Werra, Leandro
    Longpre, Shayne
    arXiv, 2023,
  • [2] Multi-Objective Linguistic Control of Large Language Models
    Dang Nguyen
    Chen, Jiuhai
    Zhou, Tianyi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 4336 - 4347
  • [3] WaveCoder: Widespread And Versatile Enhancement For Code Large Language Models By Instruction Tuning
    Yu, Zhaojian
    Zhang, Xin
    Shang, Ning
    Huang, Yangyu
    Xu, Can
    Zhao, Yishujie
    Hu, Wenxiang
    Yin, Qiufeng
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5140 - 5153
  • [4] GraphGPT: Graph Instruction Tuning for Large Language Models
    Tang, Jiabin
    Yang, Yuhao
    Wei, Wei
    Shi, Lei
    Su, Lixin
    Cheng, Suqi
    Yin, Dawei
    Huang, Chao
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 491 - 500
  • [5] InstOptima: Evolutionary Multi-objective Instruction Optimization via Large Language Model-based Instruction Operators
    Yang, Heng
    Li, Ke
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13593 - 13602
  • [6] Phased Instruction Fine-Tuning for Large Language Models
    Pang, Wei
    Zhou, Chuan
    Zhou, Xiao-Hua
    Wang, Xiaojie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5735 - 5748
  • [7] BioInstruct: instruction tuning of large language models for biomedical natural language processing
    Tran, Hieu
    Yang, Zhichao
    Yao, Zonghai
    Yu, Hong
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 1821 - 1832
  • [8] Efficient Multi-objective Genetic Tuning of Fuzzy Models for Large-Scale Regression Problems
    Casillas, Jorge
    2009 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3, 2009, : 1712 - 1717
  • [9] Multi-objective Parameter Tuning with Dynamic Compositional Surrogate Models
    Pukhkaiev, Dmytro
    Husak, Oleksandr
    Gotz, Sebastian
    Assmann, Uwe
    LEARNING AND INTELLIGENT OPTIMIZATION, LION 15, 2021, 12931 : 333 - 350
  • [10] Demystifying Instruction Mixing for Fine-tuning Large Language Models
    Wang, Renxi
    Li, Haonan
    Wu, Minghao
    Wang, Yuxia
    Han, Xudong
    Zhang, Chiyu
    Baldwin, Timothy
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 4: STUDENT RESEARCH WORKSHOP, 2024, : 86 - 93