Masked autoencoder of multi-scale convolution strategy combined with knowledge distillation for facial beauty prediction

被引:0
|
作者
Gan, Junying [1 ]
Xiong, Junling [1 ]
机构
[1] Wuyi Univ, Sch Elect Informat Engn, Jiangmen 529020, Guangdong, Peoples R China
来源
SCIENTIFIC REPORTS | 2025年 / 15卷 / 01期
基金
中国国家自然科学基金;
关键词
D O I
10.1038/s41598-025-86831-0
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Facial beauty prediction (FBP) is a leading area of research in artificial intelligence. Currently, there is a small amount of labeled data and a large amount of unlabeled data in the FBP database. The features extracted by the model based on supervised training are limited, resulting in low prediction accuracy. Masked autoencoder (MAE) is a self-supervised learning method that outperforms supervised learning methods without relying on large-scale databases. The MAE can improve the feature extraction ability of the model effectively. The multi-scale convolution strategy can expand the receptive field and combine the attention mechanism of the MAE to capture the dependency between distant pixels and acquire shallow and deep image features. Knowledge distillation can take the abundant knowledge from the teacher net to the student net, reduce the number of parameters, and compress the model. In this paper, the MAE of the multi-scale convolution strategy is combined with knowledge distillation for FBP. First, the MAE model with a multi-scale convolution strategy is constructed and used in the teacher net for pretraining. Second, the MAE model is constructed for the student net. Finally, the teacher net performs knowledge distillation, and the student net receives the loss function transmitted from the teacher net for optimization. The experimental results show that the proposed method outperforms other methods on the FBP task, improves FBP accuracy, and can be widely applied in tasks such as image classification.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Unconstrained Facial Beauty Prediction Based on Multi-scale K-Means
    GAN Junying
    ZHAI Yikui
    WANG Bin
    Chinese Journal of Electronics, 2017, 26 (03) : 548 - 556
  • [2] Optical implementation and robustness validation for multi-scale masked autoencoder
    Xue, Yizheng
    Su, Xiongfei
    Zhang, Shiyu
    Yuan, Xin
    APL PHOTONICS, 2023, 8 (04)
  • [3] Unconstrained Facial Beauty Prediction Based on Multi-scale K-Means
    Gan Junying
    Zhai Yikui
    Wang Bin
    CHINESE JOURNAL OF ELECTRONICS, 2017, 26 (03) : 548 - 556
  • [4] Multi-Scale Variational Graph AutoEncoder for Link Prediction
    Guo, Zhihao
    Wang, Feng
    Yao, Kaixuan
    Liang, Jiye
    Wang, Zhiqiang
    WSDM'22: PROCEEDINGS OF THE FIFTEENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2022, : 334 - 342
  • [5] Knowledge Fusion Distillation: Improving Distillation with Multi-scale Attention Mechanisms
    Li, Linfeng
    Su, Weixing
    Liu, Fang
    He, Maowei
    Liang, Xiaodan
    NEURAL PROCESSING LETTERS, 2023, 55 (05) : 6165 - 6180
  • [6] Knowledge Fusion Distillation: Improving Distillation with Multi-scale Attention Mechanisms
    Linfeng Li
    Weixing Su
    Fang Liu
    Maowei He
    Xiaodan Liang
    Neural Processing Letters, 2023, 55 : 6165 - 6180
  • [7] Lightweight Object Detection Combined with Multi-Scale Dilated-Convolution and Multi-Scale Deconvolution
    Yi, Qingming
    Lü, Renyi
    Shi, Min
    Luo, Aiwen
    Huanan Ligong Daxue Xuebao/Journal of South China University of Technology (Natural Science), 2022, 50 (12): : 41 - 48
  • [8] Multi-Scale Coordinate Attention Pyramid Convolution for Facial Expression Recognition
    Ni, Jinyuan
    Zhang, Jianxun
    Computer Engineering and Applications, 2023, 59 (22) : 242 - 250
  • [9] Knowledge distillation of multi-scale dense prediction transformer for self-supervised depth estimation
    Song, Jimin
    Lee, Sang Jun
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [10] Knowledge distillation of multi-scale dense prediction transformer for self-supervised depth estimation
    Jimin Song
    Sang Jun Lee
    Scientific Reports, 13