Learning from Inconsistent and Unreliable Annotators by a Gaussian Mixture Model and Bayesian Information Criterion

被引:0
|
作者
Zhang, Ping [1 ]
Obradovic, Zoran [1 ]
机构
[1] Temple Univ, Ctr Data Analyt & Biomed Informat, Philadelphia, PA 19122 USA
关键词
multiple noisy experts; data-dependent experts; Gaussian mixture model; Bayesian information criterion;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Supervised learning from multiple annotators is an increasingly important problem in machine leaning and data mining. This paper develops a probabilistic approach to this problem when annotators are not only unreliable, but also have varying performance depending on the data. The proposed approach uses a Gaussian mixture model (GMM) and Bayesian information criterion (BIC) to find the fittest model to approximate the distribution of the instances. Then the maximum a posterior (MAP) estimation of the hidden true labels and the maximum-likelihood (ML) estimation of quality of multiple annotators are provided alternately. Experiments on emotional speech classification and CASP9 protein disorder prediction tasks show performance improvement of the proposed approach as compared to the majority voting baseline and a previous data-independent approach. Moreover, the approach also provides more accurate estimates of individual annotators performance for each Gaussian component, thus paving the way for understanding the behaviors of each annotator.
引用
收藏
页码:553 / 568
页数:16
相关论文
共 50 条
  • [1] Learning from Multiple Annotators : when Data is Hard and Annotators are Unreliable
    Wolley, Chirine
    Quafafou, Mohamed
    12TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW 2012), 2012, : 514 - 521
  • [2] Improved Bayesian information criterion for mixture model selection
    Mehrjou, Arash
    Hosseini, Reshad
    Araabi, Babak Nadjar
    PATTERN RECOGNITION LETTERS, 2016, 69 : 22 - 27
  • [3] Learning from Multiple Annotators with Gaussian Processes
    Groot, Perry
    Birlutiu, Adriana
    Heskes, Tom
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2011, PT II, 2011, 6792 : 159 - +
  • [4] Comparison of speaker segmentation methods based on the Bayesian information criterion and adapted Gaussian mixture models
    Grasic, Matej
    Kos, Marko
    Zgank, Andrej
    Kacic, Zdravko
    PROCEEDINGS OF IWSSIP 2008: 15TH INTERNATIONAL CONFERENCE ON SYSTEMS, SIGNALS AND IMAGE PROCESSING, 2008, : 161 - 164
  • [5] Learning from multiple inconsistent and dependent annotators to support classification tasks
    Gil-Gonzalez, J.
    Orozco-Gutierrez, A.
    Alvarez-Meza, A.
    NEUROCOMPUTING, 2021, 423 : 236 - 247
  • [6] Complex Sequential Tasks Learning with Bayesian Inference and Gaussian Mixture Model
    Zhang, Huiwen
    Han, Xiaoning
    Zhang, Wei
    Zhou, Weijia
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO), 2018, : 1927 - 1934
  • [7] PARAMETER ESTIMATION OF GAUSSIAN MIXTURE MODEL BASED ON VARIATIONAL BAYESIAN LEARNING
    Zhao, Linchang
    Shang, Zhaowei
    Qin, Anyong
    Tang, Yuan Yan
    PROCEEDINGS OF 2018 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS (ICMLC), VOL 1, 2018, : 99 - 104
  • [8] Bayesian Repulsive Gaussian Mixture Model
    Xie, Fangzheng
    Xu, Yanxun
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2020, 115 (529) : 187 - 203
  • [9] Bayesian Correlation Filter Learning With Gaussian Scale Mixture Model for Visual Tracking
    Cao, Yuan
    Shi, Guangming
    Zhang, Tianzhu
    Dong, Weisheng
    Wu, Jinjian
    Xie, Xuemei
    Li, Xin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (05) : 3085 - 3098
  • [10] BAYESIAN LEARNING FOR HIDDEN MARKOV MODEL WITH GAUSSIAN MIXTURE STATE OBSERVATION DENSITIES
    GAUVAIN, JL
    LEE, CH
    SPEECH COMMUNICATION, 1992, 11 (2-3) : 205 - 213