-
Notifications
You must be signed in to change notification settings - Fork 98
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Help with Lead Optimization Task #179
Comments
Hi, many thanks for your interest in REINVENT and welcome to the community! When you sample from the Mol2Mol prior model you would, preferentially, receive high probability (as per NLL) SMILES sequences. You would then need to filter the resulting molecules for your objectives. You may have to sample a very large amount of compounds. Typically though, you would use Reinforcement Learning (RL, also called staged learning in REiNVENT because you can run multiple successive RL runs) using your objectives in the form of scoring functions. In this way you directly train a model to produce compounds with your desired objectives with high probability. Input examples can be found in Note that QED is a weighted sum of several physico-chemical properties including MW and TPSA. Many thanks, |
Hi Hannes, Thanks so much for your reply!
- Optimize my leads one by one with batch_size of 64, min_step of 10, max_step of 60, no max_score, and "plateau" as the termination criteria. The final molecule generated in the final step is what would be used as the "optimized version" of the lead molecule.
Thanks again! |
Hi, if your leads bind to the same target, you may want to co-optimize them. When you run RL you will generate I think the newest prior is Cheers, |
Hi Hannes, Once again, thanks so much for your help! Final question - am I correct in saying that the pretraining method for REINVENT is from Exhaustive local chemical space exploration using a transformer model by Tibo et al. and the RL method for molecular optimization is from LibINVENT: reaction-based generative scaffold decoration for in silico library design by Fialková et al. ? |
The RL algorithm has first been published in Molecular de-novo design through deep reinforcement learning. The other paper you mention is probably implemented in the Mol2Mol prior |
Hello,
Please excuse this elementary question - I am an inexperienced student on a learning journey! I am trying to compare different existing methods for lead optimization. I want to fairly compare how well different models can optimize a set of 700 DrugBank molecules using QED, MW and TPSA as objectives.
I'm unsure what configurations/tools I should use from REINVENT (since there are so many) to best accomplish this task. Is sampling using Mol2Mol sufficient?
Thank you in advance!
The text was updated successfully, but these errors were encountered: