Dominance of Bayesian Networks and Efficient Learning of Generalized Latent Class Models

被引:0
|
作者
Chen, Yixin [1 ]
Hua, Dong [1 ]
Liu, Fang [2 ]
机构
[1] Washington Univ, Dept Comp Sci, St Louis, MO 63130 USA
[2] Univ Texas Pan Amer, Dept Comp Sci, Edinburg, TX 78539 USA
关键词
D O I
10.1109/ICTAI.2008.28
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A major challenge for learning Boyesian networks is the complexity in searching the huge space of models and parameters. The computational cost is higher when the model topology is more flexible. In this paper, we propose the notion of dominance which can lead to strong pruning of the search space and significant reduction of learning complexity, and apply this notion to the Generalized Latent Class (GLC) models, a class of Bayesian networks for clustering categorical data. GLC models can address the local dependence problem in latent class analysis by assuming a very general graph structure. However The flexible topology of GLC leads to large increase of the learning complexity. We first propose the concept of dominance and related theoretical results which is general for all Bayesian networks. Based on dominance, we propose an efficient learning algorithm for GLC A core technique to prune dominated models is regularization, which can eliminate dominated models, leading to significant pruning of the search space. Significant improvements on the modeling quality and time complexity on real datasets are reported.
引用
收藏
页码:37 / +
页数:2
相关论文
共 50 条