Figure 6:Learned causal structure. 
The convergence of the algorithm is shown in Fig.7. 
 
Figure 7: The diagram of algorithm convergence. 
Both methods learned the actual causal structure 
from Fig.6. But there are significant differences in 
learning cost. According to Fig.7, 800000 samples are 
required for the traditional algorithm to learn the ac-
tual causal structure. While for the proposed algo-
rithm, only 800 samples are needed to learn the actual 
causal structure. The size of samples learned by this 
algorithm is reduced by 1000 times. From the per-
spective of learning time, the convergence time of the 
traditional algorithm is 11250s, while the conver-
gence time of the proposed algorithm is 10s. The time 
cost is reduced by 1125 times. 
It can be seen that the proposed algorithm learns 
an accurate causal structure based on the simulated 
data, and greatly outperforms the traditional algo-
rithm in terms of time cost and the size of samples, 
which verifies the advantages and effectiveness of the 
proposed algorithm. When using real data, this 
method is equally applicable and advantageous. 
4  CONCLUSION 
In order to construct an accurate causal structure and 
reduce the cost of learning, the dimension of timing 
data is reduced based on the changing features and the 
causal structure is constructed through the structural 
learning algorithms in this paper. The proposed algo-
rithm and the traditional algorithm are compared in 
learning effect and cost with an example. The pro-
posed algorithm takes the advantage of fast conver-
gence and accurate causality. However, the size of 
samples required by the proposed algorithm is still 
considerable in actual needs, so the size of samples 
required for learning needs to be further reduced in 
the next research to achieve better results. 
REFERENCES 
Adler J, Parmryd I. Quantifying colocalization by correla-
tion: The Pearson correlation coefficient is superior to 
the Mander's overlap coefficient[J]. Cytometry Part A, 
2010, 77a(8):733-742. 
Cai R, Zhang Z, Hao Z. BASSUM: A Bayesian semi-super-
vised method for classification feature selection[J]. Pat-
tern Recognition, 2011, 44(4): 811-820. 
Cai R, Zhang Z, Hao Z. Causal gene identification using 
combinatorial V-structure search[J]. Neural Networks, 
2013, 43: 63-71. 
Dong G, Wang Y, Zhang Q, et al. The methodology of Dy-
namic Uncertain Causality Graph for intelligent diagno-
sis of vertigo[J]. Computer Methods and Programs in 
Biomedicine, 2014, 113(1):162-174. 
Duan P, Yang F, Chen T, et al. Direct causality detection 
via the transfer entropy approach[J]. IEEE Transactions 
on Control Systems Technology, 2013, 21(6): 2052-
2066. 
Hao Z, Chen W, Cai R, et al. Performance optimization of 
wireless network based on canonical causal inference 
algorithm [J]. Journal of Computer Applications, 2016, 
36(8): 2114-2120. 
Hu Y, Zhang X, Ngai E W T, et al. Software project risk 
analysis using Bayesian networks with causality con-
straints[J]. Decision Support Systems, 2013, 56: 439-
449. 
Huang B, Zhang K, Lin Y, et al. Generalized score func-
tions for causal discovery[C]//Proceedings of the 24th 
ACM SIGKDD International Conference on 
Knowledge Discovery&Data Mining, 2018:1551-1560. 
Le T, Hoang T, Li J, et al. A fast PC algorithm for high 
dimensional causal discovery with multi-core PCs[J]. 
IEEE/ACM Transactions on Computational Biology 
and Bioinformatics, 2016, 16(5): 1483-1495. 
Liu M, Cai R, Hu Y, et al. Determining molecular predic-
tors of adverse drug reactions with causality analysis 
based on structure learning[J]. Journal of the American 
Medical Informatics Association, 2014, 21(2): 245-251.