Toward hardware-aware deep-learning-based dialogue systems

被引:0
|
作者
Vlad Pandelea
Edoardo Ragusa
Tom Young
Paolo Gastaldo
Erik Cambria
机构
[1] Nanyang Technological University,School of Computer Science and Engineering
[2] University of Genoa,Department of Electrical, Electronic, Telecommunications Engineering and Naval Architecture, DITEN
来源
关键词
Dialogue systems; Natural language processing; Artificial intelligence;
D O I
暂无
中图分类号
学科分类号
摘要
In the past few years, the use of transformer-based models has experienced increasing popularity as new state-of-the-art performance was achieved in several natural language processing tasks. As these models are often extremely large, however, their use for applications within embedded devices may not be feasible. In this work, we look at one such specific application, retrieval-based dialogue systems, that poses additional difficulties when deployed in environments characterized by limited resources. Research on building dialogue systems able to engage in natural sounding conversation with humans has attracted increasing attention in recent years. This has led to the rise of commercial conversational agents, such as Google Home, Alexa and Siri situated on embedded devices, that enable users to interface with a wide range of underlying functionalities in a natural and seamless manner. In part due to memory and computational power constraints, these agents necessitate frequent communication with a server in order to process the users’ queries. This communication may act as a bottleneck, resulting in delays as well as in the halt of the system should the network connection be lost or unavailable. We propose a new framework for hardware-aware retrieval-based dialogue systems based on the Dual-Encoder architecture, coupled with a clustering method to group candidates pertaining to a same conversation, that reduces storage capacity and computational power requirements.
引用
收藏
页码:10397 / 10408
页数:11
相关论文
共 50 条
  • [21] HaLo-FL: Hardware-Aware Low-Precision Federated Learning
    Venkatesha, Yeshwanth
    Bhattacharjee, Abhiroop
    Moitra, Abhishek
    Panda, Priyadarshini
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [22] A Cost-Driven Method for Deep-Learning-Based Hardware Trojan Detection
    Dong, Chen
    Yao, Yinan
    Xu, Yi
    Liu, Ximeng
    Wang, Yan
    Zhang, Hao
    Xu, Li
    SENSORS, 2023, 23 (12)
  • [23] Hardware-aware Few-shot Learning on a Memristor-based Small-world Architecture
    Raghunathan, Karthik Charan
    Demirag, Yigit
    Neftci, Emre
    Payvand, Melika
    2024 NEURO INSPIRED COMPUTATIONAL ELEMENTS CONFERENCE, NICE, 2024,
  • [24] Efficient Keyword Spotting through Hardware-Aware Conditional Execution of Deep Neural Networks
    Giraldo, J. S. P.
    O'Connor, Chris
    Verhelst, Marian
    2019 IEEE/ACS 16TH INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS (AICCSA 2019), 2019,
  • [25] Deep Quantization of Graph Neural Networks with Run-Time Hardware-Aware Training
    Hansson, Olle
    Grailoo, Mahdieh
    Gustafsson, Oscar
    Nunez-Yanez, Jose
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2024, 14553 LNCS : 33 - 47
  • [26] Toward Deep-Learning-Based Methods in Image Forgery Detection: A Survey
    Pham, Nam Thanh
    Park, Chun-Su
    IEEE ACCESS, 2023, 11 : 11224 - 11237
  • [27] Deep-Learning-Based View Interpolation Toward Improved TomoSAR Focusing
    Serafin-Garcia, Sergio Alejandro
    Nannini, Matteo
    Hansch, Ronny
    Martin-del-Campo-Becerra, Gustavo Daniel
    Reigber, Andreas
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 1
  • [28] Bayesian Deep-Learning-Based Health Prognostics Toward Prognostics Uncertainty
    Peng, Weiwen
    Ye, Zhi-Sheng
    Chen, Nan
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2020, 67 (03) : 2283 - 2293
  • [29] Hardware-aware training for large-scale and diverse deep learning inference workloads using in-memory computing-based accelerators
    Rasch, Malte J.
    Mackin, Charles
    Le Gallo, Manuel
    Chen, An
    Fasoli, Andrea
    Odermatt, Frederic
    Li, Ning
    Nandakumar, S. R.
    Narayanan, Pritish
    Tsai, Hsinyu
    Burr, Geoffrey W.
    Sebastian, Abu
    Narayanan, Vijay
    NATURE COMMUNICATIONS, 2023, 14 (01)
  • [30] Hardware-aware training for large-scale and diverse deep learning inference workloads using in-memory computing-based accelerators
    Malte J. Rasch
    Charles Mackin
    Manuel Le Gallo
    An Chen
    Andrea Fasoli
    Frédéric Odermatt
    Ning Li
    S. R. Nandakumar
    Pritish Narayanan
    Hsinyu Tsai
    Geoffrey W. Burr
    Abu Sebastian
    Vijay Narayanan
    Nature Communications, 14