Survey of Different Large Language Model Architectures: Trends, Benchmarks, and Challenges

被引:1
|
作者
Shao, Minghao [1 ]
Basit, Abdul [2 ]
Karri, Ramesh [1 ]
Shafique, Muhammad [2 ]
机构
[1] NYU, Tandon Sch Engn, New York, NY 10012 USA
[2] New York Univ Abu Dhabi, Abu Dhabi Engn Div, Abu Dhabi, U Arab Emirates
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Surveys; Transformers; Benchmark testing; Encoding; Large language models; Adaptation models; Market research; Decoding; Training; Computational modeling; Large language models (LLMs); Transformer architecture; generative models; survey; multimodal learning; deep learning; natural language processing (NLP); GENERATIVE ADVERSARIAL NETWORKS;
D O I
10.1109/ACCESS.2024.3482107
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large Language Models (LLMs) represent a class of deep learning models adept at understanding natural language and generating coherent responses to various prompts or queries. These models far exceed the complexity of conventional neural networks, often encompassing dozens of neural network layers and containing billions to trillions of parameters. They are typically trained on vast datasets, utilizing architectures based on transformer blocks. Present-day LLMs are multi-functional, capable of performing a range of tasks from text generation and language translation to question answering, as well as code generation and analysis. An advanced subset of these models, known as Multimodal Large Language Models (MLLMs), extends LLM capabilities to process and interpret multiple data modalities, including images, audio, and video. This enhancement empowers MLLMs with capabilities like video editing, image comprehension, and captioning for visual content. This survey provides a comprehensive overview of the recent advancements in LLMs. We begin by tracing the evolution of LLMs and subsequently delve into the advent and nuances of MLLMs. We analyze emerging state-of-the-art MLLMs, exploring their technical features, strengths, and limitations. Additionally, we present a comparative analysis of these models and discuss their challenges, potential limitations, and prospects for future development.
引用
收藏
页码:188664 / 188706
页数:43
相关论文
共 50 条
  • [21] A survey on large language model based autonomous agents
    WANG Lei
    MA Chen
    FENG Xueyang
    ZHANG Zeyu
    不详
    YANG Hao
    ZHANG Jingsen
    CHEN Zhiyuan
    TANG Jiakai
    CHEN Xu
    LIN Yankai
    ZHAO Wayne Xin
    WEI Zhewei
    WEN Jirong
    Frontiers of Computer Science, 2024, 18 (06)
  • [22] Trends, Challenges, and Applications of Large Language Models in Healthcare: A Bibliometric and Scoping Review
    Carchiolo, Vincenza
    Malgeri, Michele
    FUTURE INTERNET, 2025, 17 (02)
  • [23] Survey and Future Trends for FPGA Cloud Architectures
    Shahzad, Hafsah
    Sanaullah, Ahmed
    Herbordt, Martin
    2021 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2021,
  • [24] Context Aware Middleware Architectures: Survey and Challenges
    Li, Xin
    Eckert, Martina
    Martinez, Jose-Fernan
    Rubio, Gregorio
    SENSORS, 2015, 15 (08) : 20570 - 20607
  • [25] Federated Large Language Model: Solutions, Challenges and Future Directions
    Hu, Jiahui
    Wang, Dan
    Wang, Zhibo
    Pang, Xiaoyi
    Xu, Huiyu
    Ren, Ju
    Ren, Kui
    IEEE WIRELESS COMMUNICATIONS, 2024,
  • [26] Large Language Model in Critical Care Medicine: Opportunities and Challenges
    Hajijama, Sameera
    Juneja, Deven
    Nasa, Prashant
    INDIAN JOURNAL OF CRITICAL CARE MEDICINE, 2024, 28 (06) : 523 - 525
  • [27] A survey on detecting mental disorders with natural language processing: Literature review, trends and challenges
    Montejo-Raez, Arturo
    Molina-Gonzalez, M. Dolores
    Jimenez-Zafra, Salud Maria
    Garcia-Cumbreras, Miguel Angel
    Garcia-Lopez, Luis Joaquin
    COMPUTER SCIENCE REVIEW, 2024, 53
  • [28] A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions
    Huang, Lei
    Yu, Weijiang
    Ma, Weitao
    Zhong, Weihong
    Feng, Zhangyin
    Wang, Haotian
    Chen, Qianglong
    Peng, Weihua
    Feng, Xiaocheng
    Qin, Bing
    Liu, Ting
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2025, 43 (02)
  • [29] Large Language Models for Financial and Investment Management: Applications and Benchmarks
    Kong, Yaxuan
    Nie, Yuqi
    Dong, Xiaowen
    Mulvey, John M.
    Poor, H. Vincent
    Wen, Qingsong
    Zohren, Stefan
    JOURNAL OF PORTFOLIO MANAGEMENT, 2024, 51 (02): : 162 - 210
  • [30] The rise and potential of large language model based agents:a survey
    Zhiheng XI
    Wenxiang CHEN
    Xin GUO
    Wei HE
    Yiwen DING
    Boyang HONG
    Ming ZHANG
    Junzhe WANG
    Senjie JIN
    Enyu ZHOU
    Rui ZHENG
    Xiaoran FAN
    Xiao WANG
    Limao XIONG
    Yuhao ZHOU
    Weiran WANG
    Changhao JIANG
    Yicheng ZOU
    Xiangyang LIU
    Zhangyue YIN
    Shihan DOU
    Rongxiang WENG
    Wenjuan QIN
    Yongyan ZHENG
    Xipeng QIU
    Xuanjing HUANG
    Qi ZHANG
    Tao GUI
    Science China(Information Sciences), 2025, 68 (02) : 15 - 58