Learning from crowds with decision trees

被引:14
|
作者
Yang, Wenjun [1 ]
Li, Chaoqun [1 ]
Jiang, Liangxiao [2 ]
机构
[1] China Univ Geosci, Sch Math & Phys, Wuhan 430074, Peoples R China
[2] China Univ Geosci, Sch Comp Sci, Wuhan 430074, Peoples R China
关键词
Crowdsourcing learning; Weighted majority voting; Decision trees; MODEL QUALITY; STATISTICAL COMPARISONS; WEIGHTING FILTER; IMPROVING DATA; CLASSIFIERS; TOOL;
D O I
10.1007/s10115-022-01701-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Crowdsourcing systems provide an efficient way to collect labeled data by employing non-expert crowd workers. In practice, each instance obtains a multiple noisy label set from different workers. Ground truth inference algorithms are designed to infer the unknown true labels of data from multiple noisy label sets. Since there is substantial variation among different workers, evaluating the qualities of workers is crucial for ground truth inference. This paper proposes a novel algorithm called decision tree-based weighted majority voting (DTWMV). DTWMV directly takes the multiple noisy label set of each instance as its feature vector; that is, each worker is a feature of instances. Then sequential decision trees are built to calculate the weight of each feature (worker). Finally weighted majority voting is used to infer the integrated labels of instances. In DTWMV, evaluating the qualities of workers is converted to calculating the weights of features, which provides a new perspective for solving the ground truth inference problem. Then, a novel feature weight measurement based on decision trees is proposed. Our experimental results show that DTWMV can effectively evaluate the qualities of workers and improve the label quality of data.
引用
收藏
页码:2123 / 2140
页数:18
相关论文
共 50 条
  • [31] Evolutionary Learning of Interpretable Decision Trees
    Custode, Leonardo L.
    Iacca, Giovanni
    IEEE ACCESS, 2023, 11 : 6169 - 6184
  • [32] Learning Accurate and Interpretable Decision Trees
    Balcan, Maria-Florina
    Sharma, Dravyansh
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2024, 244 : 288 - 307
  • [33] Exemplar learning in fuzzy decision trees
    Janikow, CZ
    FUZZ-IEEE '96 - PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3, 1996, : 1500 - 1505
  • [34] Learning decision trees in continuous space
    Dombi, J.
    Zsiros, A.
    2001, University of Szeged (15):
  • [35] Streaming Decision Trees for Lifelong Learning
    Korycki, Lukasz
    Krawczyk, Bartosz
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, 2021, 12975 : 502 - 518
  • [36] Learning Optimal Decision Trees with SAT
    Narodytska, Nina
    Ignatiev, Alexey
    Pereira, Filipe
    Marques-Silva, Joao
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 1362 - 1368
  • [37] Learning Decision Trees for Unbalanced Data
    Cieslak, David A.
    Chawla, Nitesh V.
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PART I, PROCEEDINGS, 2008, 5211 : 241 - 256
  • [38] Learning probabilistic decision trees for AUC
    Zhang, H
    Su, J
    PATTERN RECOGNITION LETTERS, 2006, 27 (08) : 892 - 899
  • [39] STOCHASTIC INDUCTION OF DECISION TREES WITH APPLICATION TO LEARNING HAAR TREES
    Alizadeh, Azar
    Singhal, Mukesh
    Behzadan, Vahid
    Tavallali, Pooya
    Ranganath, Aditya
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 825 - 830
  • [40] Word Sense Disambiguation by Learning Decision Trees from Unlabeled Data
    Seong-Bae Park
    Byoung-Tak Zhang
    Yung Taek Kim
    Applied Intelligence, 2003, 19 : 27 - 38