A Survey of Metrics to Enhance Training Dependability in Large Language Models

被引:0
|
作者
Fang, Wenyi [1 ]
Zhang, Hao [1 ]
Gong, Ziyu [1 ]
Zeng, Longbin [1 ]
Lu, Xuhui [1 ,2 ]
Liu, Biao [1 ]
Wu, Xiaoyu [1 ]
Zheng, Yang [1 ]
Hu, Zheng [1 ]
Zhang, Xun [1 ]
机构
[1] Huawei Technol Co Ltd, Shenzhen, Peoples R China
[2] Beihang Univ, Sch Automat Sci & Elect Engn, Beijing, Peoples R China
关键词
Large Language Model; Dependability; Monitoring Metric;
D O I
10.1109/ISSREW60843.2023.00071
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The rapidly advancing field of artificial intelligence requires meticulous attention to the training and monitoring of large language models (LLMs). This paper offers a systematic analysis of existing metrics and introduces new ones, focusing on their theoretical underpinnings and practical implementations. We present empirical results and insights into the performance of selected metrics, elucidating the complex interplay of variables in the training process. Our comprehensive approach provides significant insights into LLM training, and promises to improve the dependability and efficiency of future models.
引用
收藏
页码:180 / 185
页数:6
相关论文
共 50 条
  • [1] Utilizing structural metrics from knowledge graphs to enhance the robustness quantification of large language models
    Haque, Mohd Ariful
    Kamal, Marufa
    George, Roy
    Gupta, Kishor Datta
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2024,
  • [2] Training large-scale language models with limited GPU memory: a survey
    Yu TANG
    Linbo QIAO
    Lujia YIN
    Peng LIANG
    Ao SHEN
    Zhilin YANG
    Lizhi ZHANG
    Dongsheng LI
    Frontiers of Information Technology & Electronic Engineering, 2025, 26 (03) : 309 - 331
  • [3] Training large-scale language models with limited GPU memory: a survey
    Tang, Yu
    Qiao, Linbo
    Yin, Lujia
    Liang, Peng
    Shen, Ao
    Yang, Zhilin
    Zhang, Lizhi
    Li, Dongsheng
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2025, : 309 - 331
  • [4] A survey on augmenting knowledge graphs (KGs) with large language models (LLMs): models, evaluation metrics, benchmarks, and challenges
    Ibrahim, Nourhan
    Aboulela, Samar
    Ibrahim, Ahmed
    Kashef, Rasha
    Discover Artificial Intelligence, 2024, 4 (01):
  • [5] Large Language Models in Finance: A Survey
    Li, Yinheng
    Wang, Shaofei
    Ding, Han
    Chen, Hang
    PROCEEDINGS OF THE 4TH ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2023, 2023, : 374 - 382
  • [6] Explainability for Large Language Models: A Survey
    Zhao, Haiyan
    Chen, Hanjie
    Yang, Fan
    Liu, Ninghao
    Deng, Huiqi
    Cai, Hengyi
    Wang, Shuaiqiang
    Yin, Dawei
    Du, Mengnan
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (02)
  • [7] A survey on multimodal large language models
    Yin, Shukang
    Fu, Chaoyou
    Zhao, Sirui
    Li, Ke
    Sun, Xing
    Xu, Tong
    Chen, Enhong
    NATIONAL SCIENCE REVIEW, 2024, 11 (12)
  • [8] Large language models in law: A survey
    Lai, Jinqi
    Gan, Wensheng
    Wu, Jiayang
    Qi, Zhenlian
    Yu, Philip S.
    AI OPEN, 2024, 5 : 181 - 196
  • [9] A survey of multilingual large language models
    Qin, Libo
    Chen, Qiguang
    Zhou, Yuhang
    Chen, Zhi
    Li, Yinghui
    Liao, Lizi
    Li, Min
    Che, Wanxiang
    Yu, Philip S.
    PATTERNS, 2025, 6 (01):
  • [10] A survey on LoRA of large language models
    Mao, Yuren
    Ge, Yuhang
    Fan, Yijiang
    Xu, Wenyi
    Mi, Yu
    Hu, Zhonghao
    Gao, Yunjun
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (07)