Figure 6:Learned causal structure.
The convergence of the algorithm is shown in Fig.7.
Figure 7: The diagram of algorithm convergence.
Both methods learned the actual causal structure
from Fig.6. But there are significant differences in
learning cost. According to Fig.7, 800000 samples are
required for the traditional algorithm to learn the ac-
tual causal structure. While for the proposed algo-
rithm, only 800 samples are needed to learn the actual
causal structure. The size of samples learned by this
algorithm is reduced by 1000 times. From the per-
spective of learning time, the convergence time of the
traditional algorithm is 11250s, while the conver-
gence time of the proposed algorithm is 10s. The time
cost is reduced by 1125 times.
It can be seen that the proposed algorithm learns
an accurate causal structure based on the simulated
data, and greatly outperforms the traditional algo-
rithm in terms of time cost and the size of samples,
which verifies the advantages and effectiveness of the
proposed algorithm. When using real data, this
method is equally applicable and advantageous.
4 CONCLUSION
In order to construct an accurate causal structure and
reduce the cost of learning, the dimension of timing
data is reduced based on the changing features and the
causal structure is constructed through the structural
learning algorithms in this paper. The proposed algo-
rithm and the traditional algorithm are compared in
learning effect and cost with an example. The pro-
posed algorithm takes the advantage of fast conver-
gence and accurate causality. However, the size of
samples required by the proposed algorithm is still
considerable in actual needs, so the size of samples
required for learning needs to be further reduced in
the next research to achieve better results.
REFERENCES
Adler J, Parmryd I. Quantifying colocalization by correla-
tion: The Pearson correlation coefficient is superior to
the Mander's overlap coefficient[J]. Cytometry Part A,
2010, 77a(8):733-742.
Cai R, Zhang Z, Hao Z. BASSUM: A Bayesian semi-super-
vised method for classification feature selection[J]. Pat-
tern Recognition, 2011, 44(4): 811-820.
Cai R, Zhang Z, Hao Z. Causal gene identification using
combinatorial V-structure search[J]. Neural Networks,
2013, 43: 63-71.
Dong G, Wang Y, Zhang Q, et al. The methodology of Dy-
namic Uncertain Causality Graph for intelligent diagno-
sis of vertigo[J]. Computer Methods and Programs in
Biomedicine, 2014, 113(1):162-174.
Duan P, Yang F, Chen T, et al. Direct causality detection
via the transfer entropy approach[J]. IEEE Transactions
on Control Systems Technology, 2013, 21(6): 2052-
2066.
Hao Z, Chen W, Cai R, et al. Performance optimization of
wireless network based on canonical causal inference
algorithm [J]. Journal of Computer Applications, 2016,
36(8): 2114-2120.
Hu Y, Zhang X, Ngai E W T, et al. Software project risk
analysis using Bayesian networks with causality con-
straints[J]. Decision Support Systems, 2013, 56: 439-
449.
Huang B, Zhang K, Lin Y, et al. Generalized score func-
tions for causal discovery[C]//Proceedings of the 24th
ACM SIGKDD International Conference on
Knowledge Discovery&Data Mining, 2018:1551-1560.
Le T, Hoang T, Li J, et al. A fast PC algorithm for high
dimensional causal discovery with multi-core PCs[J].
IEEE/ACM Transactions on Computational Biology
and Bioinformatics, 2016, 16(5): 1483-1495.
Liu M, Cai R, Hu Y, et al. Determining molecular predic-
tors of adverse drug reactions with causality analysis
based on structure learning[J]. Journal of the American
Medical Informatics Association, 2014, 21(2): 245-251.