Survey of Different Large Language Model Architectures: Trends, Benchmarks, and Challenges

被引:1
|
作者
Shao, Minghao [1 ]
Basit, Abdul [2 ]
Karri, Ramesh [1 ]
Shafique, Muhammad [2 ]
机构
[1] NYU, Tandon Sch Engn, New York, NY 10012 USA
[2] New York Univ Abu Dhabi, Abu Dhabi Engn Div, Abu Dhabi, U Arab Emirates
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Surveys; Transformers; Benchmark testing; Encoding; Large language models; Adaptation models; Market research; Decoding; Training; Computational modeling; Large language models (LLMs); Transformer architecture; generative models; survey; multimodal learning; deep learning; natural language processing (NLP); GENERATIVE ADVERSARIAL NETWORKS;
D O I
10.1109/ACCESS.2024.3482107
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large Language Models (LLMs) represent a class of deep learning models adept at understanding natural language and generating coherent responses to various prompts or queries. These models far exceed the complexity of conventional neural networks, often encompassing dozens of neural network layers and containing billions to trillions of parameters. They are typically trained on vast datasets, utilizing architectures based on transformer blocks. Present-day LLMs are multi-functional, capable of performing a range of tasks from text generation and language translation to question answering, as well as code generation and analysis. An advanced subset of these models, known as Multimodal Large Language Models (MLLMs), extends LLM capabilities to process and interpret multiple data modalities, including images, audio, and video. This enhancement empowers MLLMs with capabilities like video editing, image comprehension, and captioning for visual content. This survey provides a comprehensive overview of the recent advancements in LLMs. We begin by tracing the evolution of LLMs and subsequently delve into the advent and nuances of MLLMs. We analyze emerging state-of-the-art MLLMs, exploring their technical features, strengths, and limitations. Additionally, we present a comparative analysis of these models and discuss their challenges, potential limitations, and prospects for future development.
引用
收藏
页码:188664 / 188706
页数:43
相关论文
共 50 条
  • [41] When geoscience meets generative AI and large language models: Foundations, trends, and future challenges
    Hadid, Abdenour
    Chakraborty, Tanujit
    Busby, Daniel
    EXPERT SYSTEMS, 2024, 41 (10)
  • [43] Towards Large-Scale Small Object Detection: Survey and Benchmarks
    Cheng, Gong
    Yuan, Xiang
    Yao, Xiwen
    Yan, Kebing
    Zeng, Qinghua
    Xie, Xingxing
    Han, Junwei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 13467 - 13488
  • [44] English language teaching in China: Trends and challenges
    Wu, YA
    TESOL QUARTERLY, 2001, 35 (01) : 191 - 194
  • [45] Assessment of large language models for use in generative design of model based spacecraft system architectures
    Timperley, Louis Richard
    Berthoud, Lucy
    Snider, Chris
    Tryfonas, Theo
    JOURNAL OF ENGINEERING DESIGN, 2025,
  • [46] Fog Computing: Survey of Trends, Architectures, Requirements, and Research Directions
    Naha, Ranesh Kumar
    Garg, Saurabh
    Georgakopoulos, Dimitrios
    Jayaraman, Prem Prakash
    Gao, Longxiang
    Xiang, Yong
    Ranjan, Rajiv
    IEEE ACCESS, 2018, 6 : 47980 - 48009
  • [47] A survey on deep reinforcement learning architectures, applications and emerging trends
    Balhara, Surjeet
    Gupta, Nishu
    Alkhayyat, Ahmed
    Bharti, Isha
    Malik, Rami Q.
    Mahmood, Sarmad Nozad
    Abedi, Firas
    IET COMMUNICATIONS, 2022,
  • [48] A Comprehensive Survey of Recent Trends in Cloud Robotics Architectures and Applications
    Saha, Olimpiya
    Dasgupta, Prithviraj
    ROBOTICS, 2018, 7 (03)
  • [49] Agreement variation in different grammar architectures: Challenges and perspectives
    Gerasimova, Anastasia A.
    VOPROSY YAZYKOZNANIYA, 2025, (02): : 144 - 164
  • [50] Large language model integration in Philippine ophthalmology: early challenges and steps forward
    Dychiao, Robyn Gayle K.
    Alberto, Isabelle Rose I.
    Artiaga, Jose Carlo M.
    Salongcay, Recivall P.
    Celi, Leo Anthony
    LANCET DIGITAL HEALTH, 2024, 6 (05): : e308 - e308