2 Comments
Jun 25·edited Jun 25Liked by Oleg Trott, PhD

I agree with the skepticism of the author. Due to the lack of clarity on training/test bias avoidance and evaluation methods, it's not yet clear whether alphafold 3 performance in protein-ligand interaction are conclusive.

I have seen this first hand when working with CHEMBL data, a naive data-split will cause you to overestimate your model performance as (1) assay results for a particular ligand can be obtained from same/different experiments which leads to data leakage (2) many assays test for ligands with similar molecular structures, and these similar structures can end up in different train/test datasets.

One thing I would also like to see the alphafold team address is model generalization. Here is a good experiment:

1. Take your train and test ligands.

2. Compute the mean normalized jaccard index between each test ligand and train ligands.

3. Bucketize the test ligands for every 10th percentile of the normalized Jaccard Index difference.

4. Report the performance comparing alpha fold and vina.

In other words, to test model generalizability, we want to see the extent to which ligands have to be similar to training before we see gains with respect to simulation methods (e.g. vina). From experience, I expect to see some signals of overfitting.

Also, thank you for your work on vina, I learned quite a lot about this space just from reading your repo code!

Expand full comment
Jun 25·edited Jun 25Liked by Oleg Trott, PhD

Thanks Oleg for writing this piece. I feel you should just pat your back for this one. I would look at this way, despite the enormous data and computational power available to AF3 compared to what you had back then, they are “just” marginally better than what you could achieve.

Many of us owe our PhDs to your work and code!!

Thanks a million for autodock vina.

Expand full comment