Journal of Jilin University (Information Science Edition) ›› 2026, Vol. 44 ›› Issue (1): 78-86.

Previous Articles     Next Articles

Parallel Incremental Graph Bayesian Optimization for Large-Scale Virtual Screening

ZHAO Chenyang a,b , ZHAO Haishi a,b , YANG Bo a,b   

  1. a. College of Computer Science and Technology; b. Key Laboratory of Symbolic Computing and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130012, China
  • Received:2024-12-19 Online:2026-01-31 Published:2026-02-03

Abstract: Traditional methods like molecular docking often face high time costs or infeasibility in large-scale virtual screening tasks. To address this problem, a parallel graph Bayesian optimization framework incorporating incremental learning is proposed to efficiently handle such tasks. The method utilizes a deep graph Bayesian optimization framework for screening and employs parallelization to enable flexible deployment across multiple computational nodes on various servers, significantly improving computational efficiency. To tackle the issue of long surrogate model training times, an incremental learning strategy is introduced, along with an exponential moving average mechanism and a replay mechanism to mitigate catastrophic forgetting in incremental learning. Experimental results demonstrate that the framework can identify over 96% of the optimal molecules by docking only 6% of the molecular library. When deployed on four computational nodes, the parallel framework reduces time costs by 71% compared to the serial framework. With the incremental learning strategy, the total runtime is further reduced by 13. 8% , while still identifying 93. 7% of the optimal molecules. The proposed method significantly reduces the time cost of virtual screening while maintaining high screening performance. 

Key words: parallel, virtual screening, Bayesian optimization, incremental learning

CLC Number: 

  • TP391