Deep learning-powered visual place recognition for enhanced mobile multimedia communication in autonomous transport systems

被引:0
|
作者
Devi, E. M. Roopa [1 ]
Abirami, T. [2 ]
Dutta, Ashit Kumar [3 ]
Alsubai, Shtwai [1 ,2 ,3 ]
机构
[1] Kongu Engn Coll, Dept Informat Technol, Perundurai 638060, Erode, India
[2] AlMaarefa Univ, Coll Appl Sci, Dept Comp Sci & Informat Syst, Riyadh 13713, Saudi Arabia
[3] Prince Sattam bin Abdulaziz Univ, Coll Comp Engn & Sci Al Kharj, Dept Comp Sci, POB 151, Al Kharj 11942, Saudi Arabia
关键词
Visual place recognition; Autonomous transport systems; Deep learning; Hyperparameter tuning; Bilateral filtering;
D O I
10.1016/j.aej.2024.09.060
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
The progress of autonomous transport systems (ATS) involves efficient multimedia communication for real-time data tradeoffs and environmental issues. Deep learning (DL) powered visual place recognition (VPR) was developed as an effective tool to improve mobile multimedia communication in ATS. VPR relates to the capability of a method or device to recognize and identify particular places or locations from the visual scene. This procedure involves inspecting visual data, like images or video frames, to control the unique features or features connected with diverse locations. By leveraging camera sensors, VPR allows vehicles to detect their surroundings, enabling context-aware communication and enhancing the entire system's performance. DL-empowered VPR offers a transformative manner to improve mobile multimedia communication in ATS. By identifying and understanding their situation, autonomous vehicles can communicate most effectively and operate reliably and safely, paving the way for a future characterized by seamless and intelligent transportation. This article develops a novel Deep Learning-Powered Visual Place Recognition for Enhanced Multimedia Communication in Autonomous Transport Systems (DLVPR-MCATS) methodology. The main aim of the DLVPR-MCATS methodology is to recognize visual places or not utilize optimal DL approaches. For this purpose, the DLVPR-MCATS approach utilizes a bilateral filtering (BF) based preprocessing model. For the feature fusion model, the DLVPR-MCATS approach follows three models: residual network (ResNet), EfficientNet, and MobileNetv2. Moreover, the hyperparameter tuning method uses the Harris Hawks Optimization (HHO) model. Finally, the bidirectional long short-term memory (BiLSTM) technique is implemented to recognize visual places. A wide range of simulations is executed to validate the solution of the DLVPR-MCATS method. The experimental validation of the DLVPRMCATS method portrayed a superior performance over other models concerning various aspects.
引用
收藏
页码:950 / 962
页数:13
相关论文
共 50 条
  • [1] Accessible Otitis Media Screening with a Deep Learning-Powered Mobile Otoscope
    Kovvali, Omkar
    Motati, Lakshmi Sritan
    APPLICATIONS OF MEDICAL ARTIFICIAL INTELLIGENCE, AMAI 2023, 2024, 14313 : 155 - 164
  • [2] An Open Source and Open Hardware Deep Learning-powered Visual Navigation Engine for Autonomous Nano-UAVs
    Palossi, Daniele
    Conti, Francesco
    Benini, Luca
    2019 15TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING IN SENSOR SYSTEMS (DCOSS), 2019, : 604 - 611
  • [3] Deep Learning-Powered Intrusion Detection Systems: Enhancing Efficiency in Network Security
    Balamurugan, M.
    Varanasi, UshaBala
    Mangai, R. Alarmelu
    Vinayagam, P.
    Karuppaiah, S.
    Sayyed, Hifajatali
    2024 INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING, COMMUNICATION AND APPLIED INFORMATICS, ACCAI 2024, 2024,
  • [4] Deep learning-powered visual inspection for metal surfaces - Impact of annotations on algorithms based on defect characteristics
    Dubey, Pallavi
    Miller, Seth
    Gunay, Elif Elcin
    Jackman, John
    Kremer, Gul E.
    Kremer, Paul A.
    ADVANCED ENGINEERING INFORMATICS, 2024, 62
  • [5] Enhancing music recognition using deep learning-powered source separation technology for cochlear implant users
    Chang, Yuh-Jer
    Han, Ji-Yan
    Chu, Wei-Chung
    Li, Lieber Po-Hung
    Lai, Ying-Hui
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2024, 155 (03): : 1694 - 1703
  • [6] Visual place recognition: A survey from deep learning perspective
    Zhang, Xiwu
    Wang, Lei
    Su, Yan
    PATTERN RECOGNITION, 2021, 113
  • [7] Sensing Integrated DFT-Spread OFDM Waveform and Deep Learning-Powered Receiver Design for Terahertz Integrated Sensing and Communication Systems
    Wu, Yongzhi
    Lemic, Filip
    Han, Chong
    Chen, Zhi
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2023, 71 (01) : 595 - 610
  • [8] A Deep Learning-Powered TinyML Model for Gesture-Based Air Handwriting Simple Arabic Letters Recognition
    Lamaakal, Ismail
    Maleh, Yassine
    Ouahbi, Ibrahim
    El Makkaoui, Khalid
    Abd El-Latif, Ahmed A.
    DIGITAL TECHNOLOGIES AND APPLICATIONS, ICDTA 2024, VOL 4, 2024, 1101 : 32 - 42
  • [9] Deep Learning Techniques for Visual Food Recognition on a Mobile App
    De Bonis, Michele
    Amato, Giuseppe
    Falchi, Fabrizio
    Gennaro, Claudio
    Manghi, Paolo
    MULTIMEDIA AND NETWORK INFORMATION SYSTEMS, 2019, 833 : 303 - 312
  • [10] Deep learning-powered multimodal biometric authentication: integrating dynamic signatures and facial data for enhanced online security
    Salturk S.
    Kahraman N.
    Neural Computing and Applications, 2024, 36 (19) : 11311 - 11322