Fast adaptation of multi-task meta-learning for optical performance monitoring

被引:9
|
作者
Zhang, Yu [1 ,2 ]
Zhou, Peng [1 ,2 ]
Liu, Yan [1 ,2 ]
Wang, Jixiang [1 ,2 ]
Li, Chuanqi [3 ]
Lu, Ye [1 ,2 ]
机构
[1] Guangxi Normal Univ, Guilin 541004, Peoples R China
[2] Guangxi Normal Univ, Educ Dept Guangxi, Key Lab Nonlinear Circuits & Opt Commun, Guilin 541004, Peoples R China
[3] Nanning Normal Univ, Nanning 530001, Peoples R China
关键词
MODULATION FORMAT IDENTIFICATION; NEURAL-NETWORK; LOW-COMPLEXITY;
D O I
10.1364/OE.488829
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
An algorithm is proposed for few-shot-learning (FSL) jointing modulation format identification (MFI) and optical signal-to-noise ratio (OSNR) estimation. The constellation diagrams of six widely-used modulation formats over a wide range of OSNR (10-40 dB) are obtained by a dual-polarization (DP) coherent detection system at 32 GBaud. We introduce auxiliary task to model-agnostic meta-learning (MAML) which makes the gradient of meta tasks decline faster in the direction of optimal target. Ablation experiments including multi-task modeland adaptive multi-task learning (AMTL) are executed to train a data set with only 20 examples for each class. First, we discuss the impact from the number of shots and gradient descent steps for support set on the meta-learning based schemes to determine the best hyper parameters and conclude that the proposed method better captures the similarity between new and previous knowledge at 4 shot and 1 step. Withdrawn fine-tuning, the model achieves the lowest error -0.37 dB initially. Then, we simulate two other schemes (AMTL and ST-MAML), and the numerical results shows that mean square error (MSE) are -0.6 dB, -0.3 dB and -0.18 dB, respectively, proposed method has faster adaption to main task. For low order modulation formats, the proposed method almost reduces the error to 0. Meanwhile, we reveal the degree of deviation between the prediction and target and find that the deviation is mainly concentrated in the high OSNR range of 25-40 dB. Specifically, we investigate the variation curve of adaptive weights during pretraining and conclude that after 30 epoch, the model's attention was almost entirely focused on estimating OSNR. In addition, we study the generalization ability of the model by varying the transmission distance. Importantly, excellent generalization is also experimentally verified. In this paper, the method proposed will greatly reduce the cost for repetitively collecting data and the training resources required for fine-tuning models when OPM devices need to be deployed at massive nodes in dynamic optical networks.
引用
收藏
页码:23183 / 23197
页数:15
相关论文
共 50 条
  • [1] Bridging Multi-Task Learning and Meta-Learning: Towards Efficient Training and Effective Adaptation
    Wang, Haoxiang
    Zhao, Han
    Li, Bo
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [2] Multi-task metric learning for optical performance monitoring
    Zeng, Qinghui
    Lu, Ye
    Liu, Zhiqiang
    Zhang, Yu
    Li, Haiwen
    OPTICAL FIBER TECHNOLOGY, 2024, 87
  • [3] Meta-Learning for Effective Multi-task and Multilingual Modelling
    Tarunesh, Ishan
    Khyalia, Sushil
    Kumar, Vishwajeet
    Ramakrishnan, Ganesh
    Jyothi, Preethi
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 3600 - 3612
  • [4] Multi-Task Reinforcement Meta-Learning in Neural Networks
    Shakah, Ghazi
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (07) : 263 - 269
  • [5] Multi-Task and Meta-Learning with Sparse Linear Bandits
    Cella, Leonardo
    Pontil, Massimiliano
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 161, 2021, 161 : 1692 - 1702
  • [6] Revisit Multimodal Meta-Learning through the Lens of Multi-Task Learning
    Abdollahzadeh, Milad
    Malekzadeh, Touba
    Cheung, Ngai-Man
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [7] Graph Representation Learning for Multi-Task Settings: a Meta-Learning Approach
    Buffelli, Davide
    Vandin, Fabio
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [8] Homomorphisms Between Transfer, Multi-task, and Meta-learning Systems
    Cody, Tyler
    ARTIFICIAL GENERAL INTELLIGENCE, AGI 2022, 2023, 13539 : 199 - 208
  • [9] Multi-task magnetic resonance imaging reconstruction using meta-learning
    Bian, Wanyu
    Jang, Albert
    Liu, Fang
    MAGNETIC RESONANCE IMAGING, 2025, 116
  • [10] Multi-task convex combination interpolation for meta-learning with fewer tasks
    Tang, Yi
    Zhang, Liyi
    Zhang, Wuxia
    Jiang, Zuo
    KNOWLEDGE-BASED SYSTEMS, 2024, 296