Author Login Chief Editor Login Reviewer Login Editor Login Remote Office

Computer Engineering ›› 2025, Vol. 51 ›› Issue (10): 160-172. doi: 10.19678/j.issn.1000-3428.0069452

• Advanced Computing and Data Processing • Previous Articles     Next Articles

Bayesian Network Structure Learning Based on Parallel Predictive Simulated Annealing

HUANG Yun1,2, CHEN Ruoyan2, MA Li2, CAI Yiming2, LU Hengyang2, FANG Wei2,*()   

  1. 1. Jiangsu Industrial Internet Development Research Center, Suzhou 215200, Jiangsu, China
    2. School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi 214122, Jiangsu, China
  • Received:2024-02-29 Revised:2024-04-30 Online:2025-10-15 Published:2024-08-07
  • Contact: FANG Wei

基于并行预测模拟退火的贝叶斯网络结构学习

黄赟1,2, 陈若言2, 马力2, 蔡一鸣2, 陆恒杨2, 方伟2,*()   

  1. 1. 江苏省工业互联网发展研究中心,江苏 苏州 215200
    2. 江南大学人工智能与计算机学院,江苏 无锡 214122
  • 通讯作者: 方伟
  • 基金资助:
    国家自然科学基金(62073155); 国家自然科学基金(62002137); 国家自然科学基金(62106088); 国家自然科学基金(62206113)

Abstract:

Simulated Annealing (SA) is an effective method for Bayesian Network Structure Learning (BNSL). However, when handling with large-scale data, a significant search time is required. Moreover, to maintain parallel efficiency, the traditional multi-chain SA parallelization approach often requires a reduction in the number of iterations. This leads to insufficiently thorough searches when many threads are employed. Additionally, SA employs an optimal-selection update strategy during the information exchange process, which makes it prone to becoming trapped in the local optima. To address these issues, this study proposes a BNSL algorithm based on a Parallel Prediction-Based SA (PPBSA) algorithm. This algorithm ensures thoroughness in the search during the parallelization process and possesses the ability to escape local optima during the information-exchange phase. In the annealing stage of PPBSA, several generations of predicted solutions and their corresponding scores following the current solution are generated in parallel. This approach aims to guarantee search depth while substantially accelerating the search process by reducing the time spent generating and scoring subsequent solutions. When threads exchange information, a tabu list is used to restrict the search for thread solutions that have fallen into local optima, thereby enhancing the ability of the solutions to escape the local optima. Furthermore, based on the decomposability of the BDeu score, the score difference before and after perturbation in the SA process is directly calculated, significantly reducing computational redundancy. A series of experiments conducted on a set of benchmark BN compares the proposed algorithm with serial SA and other algorithms. The results demonstrate that the proposed algorithm can achieve acceleration effects of more than five times in some cases, while maintaining accuracy.

Key words: Bayesian Network (BN), structural learning, Simulated Annealing (SA), parallel algorithm, heuristic algorithm

摘要:

模拟退火(SA)是贝叶斯网络结构学习(BNSL)的有效方法,但其在大规模数据下需要耗费大量搜索时间,且传统的多链SA并行方式为保证并行效率需要减少迭代次数,导致在运行过多线程时搜索不够详尽。此外,SA在信息交换过程中使用择优更新策略,易陷入局部最优。针对上述问题,提出一种基于并行预测SA(PPBSA)的BNSL算法,其在并行化过程中确保搜索的详尽性,且在信息交换过程中具有一定的跳出局部最优的能力。PPBSA在退火阶段并行生成当前解之后的数代预测解及其评分,旨在保证搜索深度同时对搜索过程进行充分加速,减少后续多步解生成和评分计算的时间消耗。在线程交换信息时采用禁忌表对陷入局部最优的线程解进行限制搜索,提高解跳出局部最优的能力。在此基础上,基于BDeu评分的可分解性,在SA扰动过程中直接计算变动前后的评分差值,减少大量计算冗余。在一组基准BN上,将所提算法与串行SA及其他算法进行对比实验,结果表明,该算法最高可以达到5倍以上的加速效果,同时能够保证精度。

关键词: 贝叶斯网络, 结构学习, 模拟退火, 并行算法, 启发式算法