The High-performance High-throuhput Virtual Screening Framework (HHVSF) has been developed to accelerate High-Throughput Virtual Screening (HTVS) on high-performance computers. Task management and data management are two core components in HHVSF. Fine-grained computing resources are configured to support serial or threaded applications. Each task gets the input file from database through a preemptive algorithm and the failed tasks can be found and corrected. NoSQL database MongoDB is used as the data repository engine. Data is mobilized between the RAMDISK in computing node and the database. Data analysis is carried out after the computing process, and the results are stored in the database. Among the most popular molecular docking and molecular structure similarity packages, Autodock_vina (ADV) and WEGA were chosen to carry out experiments. Results show that when ADV was used for molecular docking, 10 million molecules were screened and analyzed in 22.31 h with 16000 cores, and the throughput reached up to 1324 molecules per second, averaging 145 molecules per second during the steady-running process. For WEGA, 958 million conformations were screened and analyzed in 34.12Â min with 4000 cores, of which throughput reached up to 9448 molecules per second, 6430 molecules per second on average.
CITATION STYLE
Chen, P., Yan, X., Li, J., Du, Y., & Xu, J. (2018). HHVSF: A framework to accelerate drug-based High-Throughput Virtual Screening on high-performance computers. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10776 LNCS, pp. 3–17). Springer Verlag. https://doi.org/10.1007/978-3-319-69953-0_1
Mendeley helps you to discover research relevant for your work.