A simplified adversarial architecture for cross-subject silent speech recognition using electromyography

被引:0
|
作者
Cui, Qiang [1 ,2 ,3 ]
Zhang, Xingyu [1 ,2 ,3 ]
Zhang, Yakun [1 ,2 ,3 ]
Zheng, Changyan [1 ,4 ]
Xie, Liang [1 ,2 ,3 ]
Yan, Ye [1 ,2 ,3 ]
Wu, Edmond Q. [5 ,6 ,7 ]
Yin, Erwei [1 ,2 ,3 ]
机构
[1] Acad Mil Sci AMS, Def Innovat Inst, Beijing 100071, Peoples R China
[2] Intelligent Game & Decis Lab, Beijing 100071, Peoples R China
[3] Tianjin Artificial Intelligence Innovat Ctr TAI, Tianjin 300450, Peoples R China
[4] High Tech Inst, Weifang 261000, Peoples R China
[5] Shanghai Jiao Tong Univ, Dept Automat, Shanghai 200240, Peoples R China
[6] Shanghai Jiao Tong Univ, Key Lab Syst Control & Informat Proc, Minist Educ China, Shanghai 200240, Peoples R China
[7] Shanghai Jiao Tong Univ, Shanghai Engn Res Ctr Intelligent Control & Manage, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
electromyography; silent speech recognition; cross-subject; nuclear-norm wasserstein discrepancy; domain adversarial learning; COMMUNICATION; ADAPTATION;
D O I
10.1088/1741-2552/ad7321
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective. The decline in the performance of electromyography (EMG)-based silent speech recognition is widely attributed to disparities in speech patterns, articulation habits, and individual physiology among speakers. Feature alignment by learning a discriminative network that resolves domain offsets across speakers is an effective method to address this problem. The prevailing adversarial network with a branching discriminator specializing in domain discrimination renders insufficiently direct contribution to categorical predictions of the classifier. Approach. To this end, we propose a simplified discrepancy-based adversarial network with a streamlined end-to-end structure for EMG-based cross-subject silent speech recognition. Highly aligned features across subjects are obtained by introducing a Nuclear-norm Wasserstein discrepancy metric on the back end of the classification network, which could be utilized for both classification and domain discrimination. Given the low-level and implicitly noisy nature of myoelectric signals, we devise a cascaded adaptive rectification network as the front-end feature extraction network, adaptively reshaping the intermediate feature map with automatically learnable channel-wise thresholds. The resulting features effectively filter out domain-specific information between subjects while retaining domain-invariant features critical for cross-subject recognition. Main results. A series of sentence-level classification experiments with 100 Chinese sentences demonstrate the efficacy of our method, achieving an average accuracy of 89.46% tested on 40 new subjects by training with data from 60 subjects. Especially, our method achieves a remarkable 10.07% improvement compared to the state-of-the-art model when tested on 10 new subjects with 20 subjects employed for training, surpassing its result even with three times training subjects. Significance. Our study demonstrates an improved classification performance of the proposed adversarial architecture using cross-subject myoelectric signals, providing a promising prospect for EMG-based speech interactive application.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] Joint EEG Feature Transfer and Semisupervised Cross-Subject Emotion Recognition
    Peng, Yong
    Liu, Honggang
    Kong, Wanzeng
    Nie, Feiping
    Lu, Bao-Liang
    Cichocki, Andrzej
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (07) : 8104 - 8115
  • [42] Comprehensive Multisource Learning Network for Cross-Subject Multimodal Emotion Recognition
    Chen, Chuangquan
    Li, Zhencheng
    Kou, Kit Ian
    Du, Jie
    Li, Chen
    Wang, Hongtao
    Vong, Chi-Man
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025, 9 (01): : 365 - 380
  • [43] Hybrid transfer learning strategy for cross-subject EEG emotion recognition
    Lu, Wei
    Liu, Haiyan
    Ma, Hua
    Tan, Tien-Ping
    Xia, Lingnan
    FRONTIERS IN HUMAN NEUROSCIENCE, 2023, 17
  • [44] Transfer Learning Enhanced Cross-Subject Hand Gesture Recognition with sEMG
    Shenyilang Zhang
    Yinfeng Fang
    Jiacheng Wan
    Guozhang Jiang
    Gongfa Li
    Journal of Medical and Biological Engineering, 2023, 43 : 672 - 688
  • [45] Cross-Subject and Cross-Modal Transfer for Generalized Abnormal Gait Pattern Recognition
    Gu, Xiao
    Guo, Yao
    Deligianni, Fani
    Lo, Benny
    Yang, Guang-Zhong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (02) : 546 - 560
  • [46] FMLAN: A novel framework for cross-subject and cross-session EEG emotion recognition
    Yu, Peng
    He, Xiaopeng
    Li, Haoyu
    Dou, Haowen
    Tan, Yeyu
    Wu, Hao
    Chen, Badong
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 100
  • [47] Distilling Invariant Representations With Domain Adversarial Learning for Cross-Subject Children Seizure Prediction
    Zhang, Ziye
    Liu, Aiping
    Gao, Yikai
    Cui, Xinrui
    Qian, Ruobing
    Chen, Xun
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (01) : 202 - 211
  • [48] Standardization-refinement domain adaptation method for cross-subject EEG-based classification in imagined speech recognition
    Jimenez-Guarneros, Magdiel
    Gomez-Gil, Pilar
    PATTERN RECOGNITION LETTERS, 2021, 141 : 54 - 60
  • [49] Cross-Subject Emotion Recognition From Multichannel EEG Signals Using Multivariate Decomposition and Ensemble Learning
    Vempati, Raveendrababu
    Sharma, Lakhan Dev
    Tripathy, Rajesh Kumar
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2025, 17 (01) : 77 - 88
  • [50] Cross-subject EEG emotion recognition using multi-source domain manifold feature selection
    She, Qingshan
    Shi, Xinsheng
    Fang, Feng
    Ma, Yuliang
    Zhang, Yingchun
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 159