Toward Low-Resource Languages Machine Translation: A Language-Specific Fine-Tuning With LoRA for Specialized Large Language Models

被引:0
|
作者
Liang, Xiao [1 ,2 ]
Khaw, Yen-Min Jasmina [1 ]
Liew, Soung-Yue [3 ]
Tan, Tien-Ping [4 ]
Qin, Donghong [2 ]
机构
[1] Univ Tunku Abdul Rahman, Fac Informat & Commun Technol, Dept Comp Sci, Kampar 31900, Malaysia
[2] Guangxi Minzu Univ, Sch Artificial Intelligence, Nanning 530008, Peoples R China
[3] Univ Tunku Abdul Rahman, Fac Informat & Commun Technol, Dept Comp & Commun Technol, Kampar 31900, Malaysia
[4] Univ Sains Malaysia, Sch Comp Sci, George Town 11700, Malaysia
来源
IEEE ACCESS | 2025年 / 13卷
关键词
Machine translation; low-resource languages; large language models; parameter-efficient fine-tuning; LoRA;
D O I
10.1109/ACCESS.2025.3549795
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the field of computational linguistics, addressing machine translation (MT) challenges for low-resource languages remains crucial, as these languages often lack extensive data compared to high-resource languages. General large language models (LLMs), such as GPT-4 and Llama, primarily trained on monolingual corpora, face significant challenges in translating low-resource languages, often resulting in subpar translation quality. This study introduces Language-Specific Fine-Tuning with Low-rank adaptation (LSFTL), a method that enhances translation for low-resource languages by optimizing the multi-head attention and feed-forward networks of Transformer layers through low-rank matrix adaptation. LSFTL preserves the majority of the model parameters while selectively fine-tuning key components, thereby maintaining stability and enhancing translation quality. Experiments on non-English centered low-resource Asian languages demonstrated that LSFTL improved COMET scores by 1-3 points compared to specialized multilingual machine translation models. Additionally, LSFTL's parameter-efficient approach allows smaller models to achieve performance comparable to their larger counterparts, highlighting its significance in making machine translation systems more accessible and effective for low-resource languages.
引用
收藏
页码:46616 / 46626
页数:11
相关论文
共 50 条
  • [31] Repeatability of Fine-Tuning Large Language Models Illustrated Using QLoRA
    Alahmari, Saeed S.
    Hall, Lawrence O.
    Mouton, Peter R.
    Goldgof, Dmitry B.
    IEEE ACCESS, 2024, 12 : 153221 - 153231
  • [32] Fine-tuning large language models for rare disease concept normalization
    Wang, Andy
    Liu, Cong
    Yang, Jingye
    Weng, Chunhua
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 2076 - 2083
  • [33] How to choose the best pivot language for automatic translation of low-resource languages
    Paul, Michael
    Finch, Andrew
    Sumita, Eiichrio
    ACM Transactions on Asian Language Information Processing, 2013, 12 (04):
  • [34] Morph-fitting: Fine-Tuning Word Vector Spaces with Simple Language-Specific Rules
    Vulic, Ivan
    Mrksic, Nikola
    Reichart, Roi
    Seaghdha, Diarmuid O.
    Young, Steve
    Korhonen, Anna
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 56 - 68
  • [35] Data Augmentation with Translation Memories for Desktop Machine Translation Fine-tuning in 3 Language Pairs
    Dogru, Gokhan
    Moorkens, Joss
    JOURNAL OF SPECIALISED TRANSLATION, 2024, (41): : 149 - 178
  • [36] Integrating Task Specific Information into Pretrained Language Models for Low Resource Fine Tuning
    Wang, Rui
    Si, Shijing
    Wang, Guoyin
    Zhang, Lei
    Carin, Lawrence
    Henao, Ricardo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020,
  • [37] The Task of Post-Editing Machine Translation for the Low-Resource Language
    Rakhimova, Diana
    Karibayeva, Aidana
    Turarbek, Assem
    APPLIED SCIENCES-BASEL, 2024, 14 (02):
  • [38] Unsupervised Multimodal Machine Translation for Low-resource Distant Language Pairs
    Tayir, Turghun
    Li, Lin
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (04)
  • [39] Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
    Zhang, Luoming
    Lou, Zhenyu
    Ying, Yangwei
    Yang, Cheng
    Zhou, Hong
    APPLIED SCIENCES-BASEL, 2025, 15 (01):
  • [40] Exploring Large Language Models for Low-Resource IT Information Extraction
    Bhavya, Bhavya
    Isaza, Paulina Toro
    Deng, Yu
    Nidd, Michael
    Azad, Amar Prakash
    Shwartz, Larisa
    Zhai, ChengXiang
    2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 1203 - 1212