ArithmeticGPT: empowering small-size large language models with advanced arithmetic skills

被引:0
|
作者
Liu, Zitao
Zheng, Ying
Yin, Zhibo
Chen, Jiahao
Liu, Tianqiao
Tian, Mi
Luo, Weiqi
机构
基金
国家重点研发计划;
关键词
Large language models; Problem-solving; Math reasoning; Curriculum learning;
D O I
10.1007/s10994-024-06681-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) have shown remarkable capabilities in understanding and generating language across a wide range of domains. However, their performance in advanced arithmetic calculation remains a significant challenge, especially for small-size LLMs. Therefore, in this paper, we propose ArithmeticGPT, a practical framework designed to enhance the advanced arithmetic skills for small-size LLMs. We carefully curate an arithmetic instruction dataset, ArithInstruct, that is able to teach the small-size LLMs to trigger a self-developed internal calculation API for precise computations without explicit instructions. The advanced arithmetic calculation results are seamlessly generated within natural language sentences. Furthermore, we empirically design a practical three-stage strategy for fine-tuning the small-size LLMs with ArithInstruct to enable the advanced arithmetic skills and keep the models' original abilities such as commonsense reasoning and question answering. We evaluate ArithmeticGPT on six public math related datasets with 17 state-of-the-art LLM baselines and experimental results demonstrate the superiority of our approach. To encourage reproducible research, we make our data and code publicly available at https://github.com/ai4ed/ArithmeticGPT.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] AtomTool: Empowering Large Language Models with Tool Utilization Skills
    Li, Yongle
    Zhang, Zheng
    Zhang, Junqi
    Hu, Wenbo
    Wu, Yongyu
    Hong, Richang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT 1, 2025, 15031 : 323 - 337
  • [2] PEER: Empowering Writing with Large Language Models
    Sessler, Kathrin
    Xiang, Tao
    Bogenrieder, Lukas
    Kasneci, Enkelejda
    RESPONSIVE AND SUSTAINABLE EDUCATIONAL FUTURES, EC-TEL 2023, 2023, 14200 : 755 - 761
  • [3] Small-Size vs Large-Size Burr for Rotational Atherectomy
    Levi, Yaniv
    Lavi, Shahar
    Solomonica, Amir
    Israeli, Zeev
    Bagur, Rodrigo
    JOURNAL OF INVASIVE CARDIOLOGY, 2019, 31 (06): : 183 - 186
  • [4] Empowering Large Language Models for Textual Data Augmentation
    Li, Yichuan
    Ding, Kaize
    Wang, Jianling
    Lee, Kyumin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 12734 - 12751
  • [5] chatHPC: Empowering HPC users with large language models
    Yin, Junqi
    Hines, Jesse
    Herron, Emily
    Ghosal, Tirthankar
    Liu, Hong
    Prentice, Suzanne
    Lama, Vanessa
    Wang, Feiyi
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (01):
  • [6] Empowering Time Series Analysis with Large Language Models: A Survey
    Jiang, Yushan
    Pan, Zijie
    Zhang, Xikun
    Garg, Sahil
    Schneider, Anderson
    Nevmyvaka, Yuriy
    Song, Dongjin
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 8095 - 8103
  • [7] PointLLM: Empowering Large Language Models to Understand Point Clouds
    Xu, Runsen
    Wang, Xiaolong
    Wang, Tai
    Chen, Yilun
    Pang, Jiangmiao
    Lin, Dahua
    COMPUTER VISION - ECCV 2024, PT XXV, 2025, 15083 : 131 - 147
  • [8] Performance evaluation of large-size tilting-pad thrust bearings based on results from small-size models
    Wodtke, Michal
    Wasilczuk, Michal
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART J-JOURNAL OF ENGINEERING TRIBOLOGY, 2025,
  • [9] OTSI OF LOWER-HYBRID WAVES IN LARGE-SIZE AND SMALL-SIZE DEVICES
    TRIPATHI, VK
    PRASAD, PVSR
    JOURNAL OF PLASMA PHYSICS, 1989, 41 : 13 - 22
  • [10] A scheme of hiding large-size image into small-size image based on FCdDNet
    Liu, Lianshan
    Tang, Li
    Tong, Shanshan
    Huang, Yu
    PEERJ COMPUTER SCIENCE, 2024, 10