Unknown

Dataset Information

0

Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark.


ABSTRACT: The superior performance of machine-learning scoring functions for docking has caused a series of debates on whether it is due to learning knowledge from training data that are similar in some sense to the test data. With a systematically revised methodology and a blind benchmark realistically mimicking the process of prospective prediction of binding affinity, we have evaluated three broadly used classical scoring functions and five machine-learning counterparts calibrated with both random forest and extreme gradient boosting using both solo and hybrid features, showing for the first time that machine-learning scoring functions trained exclusively on a proportion of as low as 8% complexes dissimilar to the test set already outperform classical scoring functions, a percentage that is far lower than what has been recently reported on all the three CASF benchmarks. The performance of machine-learning scoring functions is underestimated due to the absence of similar samples in some artificially created training sets that discard the full spectrum of complexes to be found in a prospective environment. Given the inevitability of any degree of similarity contained in a large dataset, the criteria for scoring function selection depend on which one can make the best use of all available materials. Software code and data are provided at https://github.com/cusdulab/MLSF for interested readers to rapidly rebuild the scoring functions and reproduce our results, even to make extended analyses on their own benchmarks.

SUBMITTER: Li H 

PROVIDER: S-EPMC8575004 | biostudies-literature | 2021 Nov

REPOSITORIES: biostudies-literature

altmetric image

Publications

Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark.

Li Hongjian H   Lu Gang G   Sze Kam-Heung KH   Su Xianwei X   Chan Wai-Yee WY   Leung Kwong-Sak KS  

Briefings in bioinformatics 20211101 6


The superior performance of machine-learning scoring functions for docking has caused a series of debates on whether it is due to learning knowledge from training data that are similar in some sense to the test data. With a systematically revised methodology and a blind benchmark realistically mimicking the process of prospective prediction of binding affinity, we have evaluated three broadly used classical scoring functions and five machine-learning counterparts calibrated with both random fore  ...[more]

Similar Datasets

| S-EPMC5555960 | biostudies-literature
| S-EPMC11832058 | biostudies-literature
| S-EPMC9020343 | biostudies-literature
| S-EPMC10335560 | biostudies-literature
| S-EPMC10297005 | biostudies-literature
| S-EPMC10688188 | biostudies-literature
| PRJEB33197 | ENA
| S-EPMC3190652 | biostudies-literature
| S-EPMC8061123 | biostudies-literature
| S-EPMC11783320 | biostudies-literature