摘要
It is unpractical to learn the optimal structure of a big Bayesian network(BN)by exhausting the feasible structures,since the number of feasible structures is super exponential on the number of nodes.This paper proposes an approach to layer nodes of a BN by using the conditional independence testing.The parents of a node layer only belong to the layer,or layers who have priority over the layer.When a set of nodes has been layered,the number of feasible structures over the nodes can be remarkably reduced,which makes it possible to learn optimal BN structures for bigger sizes of nodes by accurate algorithms.Integrating the dynamic programming(DP)algorithm with the layering approach,we propose a hybrid algorithm—layered optimal learning(LOL)to learn BN structures.Benefitted by the layering approach,the complexity of the DP algorithm reduces to O(ρ2^n?1)from O(n2^n?1),whereρ<n.Meanwhile,the memory requirements for storing intermediate results are limited to O(C k#/k#^2 )from O(Cn/n^2 ),where k#<n.A case study on learning a standard BN with 50 nodes is conducted.The results demonstrate the superiority of the LOL algorithm,with respect to the Bayesian information criterion(BIC)score criterion,over the hill-climbing,max-min hill-climbing,PC,and three-phrase dependency analysis algorithms.
It is unpractical to learn the optimal structure of a big Bayesian network(BN) by exhausting the feasible structures, since the number of feasible structures is super exponential on the number of nodes. This paper proposes an approach to layer nodes of a BN by using the conditional independence testing. The parents of a node layer only belong to the layer, or layers who have priority over the layer. When a set of nodes has been layered, the number of feasible structures over the nodes can be remarkably reduced,which makes it possible to learn optimal BN structures for bigger sizes of nodes by accurate algorithms. Integrating the dynamic programming(DP) algorithm with the layering approach, we propose a hybrid algorithm — layered optimal learning(LOL) to learn cρoaB fta Ne<ts h ers enetrs.s tuDuuMclPttdesu y ara aleongrsnweo. lrhlieBiitlameheri,nmt nteei hrnfidete gtt deomau d e cOsbtem(aysC on ttrkdhkoy2#ae #r rOled)a( qyfρBure2 ioNrrin emnw-mgi1 O te)ah( n pftC5 rpso0 n r2 n fomon)ar,o cOsdwht(e,oh nsrte i2 hirnsnee g c-ckio1 no#nt)m, ed r<puwlm cehntexee.id rtdiAey-.The results demonstrate the superiority of the LOL algorithm, with respect to the Bayesian information criterion(BIC) score criterion,over the hill-climbing, max-min hill-climbing, PC, and three-phrase dependency analysis algorithms.
基金
supported by the National Natural Science Foundation of China(61573285)