On The Difficulty of Validating Molecular Generative Models Realistically: A Case Study on Public and Proprietary Data

20 November 2023, Version 2
This content is a preprint and has not undergone peer review at the time of posting.

Abstract

While a multitude of deep generative models have recently emerged there exists no best practice for their practically relevant validation. On the one hand, novel de novo-generated molecules cannot be refuted by retrospective validation (so that this type of validation is biased); but on the other hand prospective validation is expensive and then often biased by the human selection process. In this case study, we frame retrospective validation as the ability to mimic human drug design, by answering the following question: Can a generative model trained on early-stage project compounds generate middle/late-stage compounds de novo? To this end, we used experimental data that contains the elapsed time of a synthetic expansion following hit identification from five public (where the time series was pre-processed to better reflect realistic synthetic expansions) and six in-house project datasets, and used REINVENT as a widely adopted RNN-based generative model. After splitting the dataset and training REINVENT on early-stage compounds, we found that rediscovery of middle/late-stage compounds was much higher in public projects (at 1.60%, 0.64%, and 0.21% of the top 100, 500, and 5,000 scored generated compounds) than in in-house projects (where the values were 0.00%, 0.03%, and 0.04%, respectively). Similarly, average single nearest neighbour similarity between early- and middle/late-stage compounds in public projects was higher between active compounds than inactive compounds; however, for in-house projects the converse was true, which makes rediscovery (if so desired) more difficult. We hence show that the generative model recovers very few middle/late-stage compounds from real-world drug discovery projects, highlighting the fundamental difference between purely algorithmic design and drug discovery as a real-world process. Evaluating de novo compound design approaches appears, based on the current study, difficult or even impossible to do retrospectively. "Scientific Contribution" This contribution hence illustrates aspects of evaluating the performance of generative models in a real-world setting which have not been extensively described previously and which hopefully contribute to their further future development.

Keywords

Artificial Intelligence
AI
Ligand-based drug design
LBDD
Generative models
De novo molecular generation
Recurrent neural network
REINVENT
Reinforcement learning
Validation

Supplementary materials

Title
Description
Actions
Title
Supporitng Information
Description
Supporitng Information (Figures and Tables)
Actions
Title
Supporting_Dataset_1_DRD2
Description
Supporting_Dataset_1_DRD2
Actions
Title
Supporting_Dataset_2_GSK3
Description
Supporting_Dataset_2_GSK3
Actions
Title
Supporting_Dataset_3_CDK2
Description
Supporting_Dataset_3_CDK2
Actions
Title
Supporting_Dataset_4_EGFR
Description
Supporting_Dataset_4_EGFR
Actions
Title
Supporting_Dataset_5_ADRB2
Description
Supporting_Dataset_5_ADRB2
Actions

Comments

Comments are not moderated before they are posted, but they can be removed by the site moderators if they are found to be in contravention of our Commenting Policy [opens in a new tab] - please read this policy before you post. Comments should be used for scholarly discussion of the content in question. You can find more information about how to use the commenting feature here [opens in a new tab] .
This site is protected by reCAPTCHA and the Google Privacy Policy [opens in a new tab] and Terms of Service [opens in a new tab] apply.