Optimization of Molecules via Deep Reinforcement Learning
Zhenpeng Zhou, Steven Kearnes, Li Li, Richard N. Zare, Patrick Riley
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/google-research/google-research/tree/master/mol_dqnOfficialtf★ 0
- github.com/danilonumeroso/MEGpytorch★ 17
- github.com/junyoung0131/Mol-DQNtf★ 0
- github.com/tangxiangru/RL-for-RNA-designtf★ 0
- github.com/aksub99/MolDQN-pytorchpytorch★ 0
- github.com/caiyingchun/MolDQNtf★ 0
- github.com/2023-MindSpore-4/Code12/tree/main/d2l/chapter_11_optimizationmindspore★ 0
Abstract
We present a framework, which we call Molecule Deep Q-Networks (MolDQN), for molecule optimization by combining domain knowledge of chemistry and state-of-the-art reinforcement learning techniques (double Q-learning and randomized value functions). We directly define modifications on molecules, thereby ensuring 100\% chemical validity. Further, we operate without pre-training on any dataset to avoid possible bias from the choice of that set. Inspired by problems faced during medicinal chemistry lead optimization, we extend our model with multi-objective reinforcement learning, which maximizes drug-likeness while maintaining similarity to the original molecule. We further show the path through chemical space to achieve optimization for a molecule to understand how the model works.