Kids' Emotion Recognition Using Various Deep-Learning Models with Explainable AI

被引:10
|
作者
Rathod, Manish [1 ]
Dalvi, Chirag [1 ]
Kaur, Kulveen [1 ]
Patil, Shruti [2 ]
Gite, Shilpa [2 ]
Kamat, Pooja [1 ]
Kotecha, Ketan [2 ]
Abraham, Ajith [3 ]
Gabralla, Lubna Abdelkareim [4 ]
机构
[1] Deemed Univ, Symbiosis Int Univ, Symbiosis Ctr Appl Artificial Intelligence SCAAI, Pune 412115, Maharashtra, India
[2] Deemed Univ, Symbiosis Int Univ, Symbiosis Inst Technol, Comp Sci & Informat Technol Dept, Pune 412115, Maharashtra, India
[3] Machine Intelligence Res Labs MIR Labs, Auburn, WA 98071 USA
[4] Princess Nourah Bint Abdulrahman Univ, Coll Appl, Dept Comp Sci & Informat Technol, Riyadh 11671, Saudi Arabia
关键词
kids' emotion recognition; FER; explainable artificial intelligence; LIRIS; children emotion dataset; online learning;
D O I
10.3390/s22208066
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Human ideas and sentiments are mirrored in facial expressions. They give the spectator a plethora of social cues, such as the viewer's focus of attention, intention, motivation, and mood, which can help develop better interactive solutions in online platforms. This could be helpful for children while teaching them, which could help in cultivating a better interactive connect between teachers and students, since there is an increasing trend toward the online education platform due to the COVID-19 pandemic. To solve this, the authors proposed kids' emotion recognition based on visual cues in this research with a justified reasoning model of explainable AI. The authors used two datasets to work on this problem; the first is the LIRIS Children Spontaneous Facial Expression Video Database, and the second is an author-created novel dataset of emotions displayed by children aged 7 to 10. The authors identified that the LIRIS dataset has achieved only 75% accuracy, and no study has worked further on this dataset in which the authors have achieved the highest accuracy of 89.31% and, in the authors' dataset, an accuracy of 90.98%. The authors also realized that the face construction of children and adults is different, and the way children show emotions is very different and does not always follow the same way of facial expression for a specific emotion as compared with adults. Hence, the authors used 3D 468 landmark points and created two separate versions of the dataset from the original selected datasets, which are LIRIS-Mesh and Authors-Mesh. In total, all four types of datasets were used, namely LIRIS, the authors' dataset, LIRIS-Mesh, and Authors-Mesh, and a comparative analysis was performed by using seven different CNN models. The authors not only compared all dataset types used on different CNN models but also explained for every type of CNN used on every specific dataset type how test images are perceived by the deep-learning models by using explainable artificial intelligence (XAI), which helps in localizing features contributing to particular emotions. The authors used three methods of XAI, namely Grad-CAM, Grad-CAM++, and SoftGrad, which help users further establish the appropriate reason for emotion detection by knowing the contribution of its features in it.
引用
收藏
页数:21
相关论文
共 50 条
  • [31] An explainable fast deep neural network for emotion recognition
    Di Luzio, Francesco
    Rosato, Antonello
    Panella, Massimo
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 100
  • [32] Explainable AI for Bearing Fault Prognosis Using Deep Learning Techniques
    Sanakkayala, Deva Chaitanya
    Varadarajan, Vijayakumar
    Kumar, Namya
    Karan
    Soni, Girija
    Kamat, Pooja
    Kumar, Satish
    Patil, Shruti
    Kotecha, Ketan
    MICROMACHINES, 2022, 13 (09)
  • [33] Quality of Pre-trained Deep-Learning Models for Palmprint Recognition
    Rosca, Valentin
    Ignat, Anca
    2020 22ND INTERNATIONAL SYMPOSIUM ON SYMBOLIC AND NUMERIC ALGORITHMS FOR SCIENTIFIC COMPUTING (SYNASC 2020), 2020, : 202 - 209
  • [34] A novel framework of multiclass skin lesion recognition from dermoscopic images using deep learning and explainable AI
    Ahmad, Naveed
    Shah, Jamal Hussain
    Khan, Muhammad Attique
    Baili, Jamel
    Ansari, Ghulam Jillani
    Tariq, Usman
    Kim, Ye Jin
    Cha, Jae-Hyuk
    FRONTIERS IN ONCOLOGY, 2023, 13
  • [35] Factors in Emotion Recognition With Deep Learning Models Using Speech and Text on Multiple Corpora
    Braunschweiler, Norbert
    Doddipatla, Rama
    Keizer, Simon
    Stoyanchev, Svetlana
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 722 - 726
  • [36] Automatic Recognition of Laryngoscopic Images Using a Deep-Learning Technique
    Ren, Jianjun
    Jing, Xueping
    Wang, Jing
    Ren, Xue
    Xu, Yang
    Yang, Qiuyun
    Ma, Lanzhi
    Sun, Yi
    Xu, Wei
    Yang, Ning
    Zou, Jian
    Zheng, Yongbo
    Chen, Min
    Gan, Weigang
    Xiang, Ting
    An, Junnan
    Liu, Ruiqing
    Lv, Cao
    Lin, Ken
    Zheng, Xianfeng
    Lou, Fan
    Rao, Yufang
    Yang, Hui
    Liu, Kai
    Liu, Geoffrey
    Lu, Tao
    Zheng, Xiujuan
    Zhao, Yu
    LARYNGOSCOPE, 2020, 130 (11): : E686 - E693
  • [37] Multimodal Emotion Recognition using Deep Learning Architectures
    Ranganathan, Hiranmayi
    Chakraborty, Shayok
    Panchanathan, Sethuraman
    2016 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2016), 2016,
  • [38] Multimodal Arabic emotion recognition using deep learning
    Al Roken, Noora
    Barlas, Gerassimos
    SPEECH COMMUNICATION, 2023, 155
  • [39] Evaluating Explainable AI Methods in Deep Learning Models for Early Detection of Cerebral Palsy
    Pellano, Kimji N.
    Strumke, Inga
    Groos, Daniel
    Adde, Lars
    Ihlen, Espen F. Alexander
    IEEE ACCESS, 2025, 13 : 10126 - 10138
  • [40] Explainable feature selection and deep learning based emotion recognition in virtual reality using eye tracker and physiological data
    Alharbi, Hadeel
    FRONTIERS IN MEDICINE, 2024, 11