ArithmeticGPT: empowering small-size large language models with advanced arithmetic skills

被引:0
|
作者
Liu, Zitao
Zheng, Ying
Yin, Zhibo
Chen, Jiahao
Liu, Tianqiao
Tian, Mi
Luo, Weiqi
机构
基金
国家重点研发计划;
关键词
Large language models; Problem-solving; Math reasoning; Curriculum learning;
D O I
10.1007/s10994-024-06681-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) have shown remarkable capabilities in understanding and generating language across a wide range of domains. However, their performance in advanced arithmetic calculation remains a significant challenge, especially for small-size LLMs. Therefore, in this paper, we propose ArithmeticGPT, a practical framework designed to enhance the advanced arithmetic skills for small-size LLMs. We carefully curate an arithmetic instruction dataset, ArithInstruct, that is able to teach the small-size LLMs to trigger a self-developed internal calculation API for precise computations without explicit instructions. The advanced arithmetic calculation results are seamlessly generated within natural language sentences. Furthermore, we empirically design a practical three-stage strategy for fine-tuning the small-size LLMs with ArithInstruct to enable the advanced arithmetic skills and keep the models' original abilities such as commonsense reasoning and question answering. We evaluate ArithmeticGPT on six public math related datasets with 17 state-of-the-art LLM baselines and experimental results demonstrate the superiority of our approach. To encourage reproducible research, we make our data and code publicly available at https://github.com/ai4ed/ArithmeticGPT.
引用
收藏
页数:23
相关论文
共 50 条
  • [41] Comuniqa : Exploring Large Language Models for Improving English Speaking Skills
    Mhasakar, Manas
    Sharma, Shikhar
    Mehra, Apurv
    Venaik, Utkarsh
    Singhal, Ujjwal
    Kumar, Dhruv
    Mittal, Kashish
    PROCEEDINGS OF THE ACM SIGCAS/SIGCHI CONFERENCE ON COMPUTING AND SUSTAINABLE SOCIETIES 2024, COMPASS 2024, 2024, : 256 - 267
  • [42] Some design trade-offs for large CNN chips using small-size transistors
    RodriguezVazquez, A
    Linan, G
    DominguezCastro, R
    Huertas, JL
    Espejo, S
    ISCAS '97 - PROCEEDINGS OF 1997 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOLS I - IV: CIRCUITS AND SYSTEMS IN THE INFORMATION AGE, 1997, : 741 - 744
  • [44] Empowering Education with Intelligent Systems: Exploring Large Language Models and the NAO Robot for Information Retrieval
    Fragakis, Nikos
    Trichopoulos, Georgios
    Caridakis, George
    ELECTRONICS, 2025, 14 (06):
  • [45] Empowering Psychotherapy with Large Language Models: Cognitive Distortion Detection through Diagnosis of Thought Prompting
    Chen, Zhiyu
    Lu, Yujie
    Wang, William Yang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4295 - 4304
  • [46] Empowering Molecule Discovery for Molecule-Caption Translation With Large Language Models: A ChatGPT Perspective
    Li, Jiatong
    Liu, Yunqing
    Fan, Wenqi
    Wei, Xiao-Yong
    Liu, Hui
    Tang, Jiliang
    Li, Qing
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (11) : 6071 - 6083
  • [47] LOGIC-LM: Empowering Large Language Models with Symbolic Solvers for Faithful Logical Reasoning
    Pan, Liangming
    Albalak, Alon
    Wang, Xinyi
    Wang, William Yang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3806 - 3824
  • [48] Empowering Few-Shot Recommender Systems With Large Language Models-Enhanced Representations
    Wang, Zhoumeng
    IEEE ACCESS, 2024, 12 : 29144 - 29153
  • [49] Empowering Large Language Models to Leverage Domain-Specific Knowledge in E-Learning
    Lu, Ruei-Shan
    Lin, Ching-Chang
    Tsao, Hsiu-Yuan
    APPLIED SCIENCES-BASEL, 2024, 14 (12):
  • [50] Small, Medium, and Large Language Models for Text-to-SQL
    Oliveira, Aiko
    Nascimento, Eduardo
    Pinheiro, Joao
    Avila, Caio Viktor S.
    Coelho, Gustavo
    Feijo, Lucas
    Izquierdo, Yenier
    Garcia, Grettel
    Paes Leme, Luiz Andre P.
    Lemos, Melissa
    Casanova, Marco A.
    CONCEPTUAL MODELING, ER 2024, 2025, 15238 : 276 - 294