Skip navigation

DSpace

機構典藏 DSpace 系統致力於保存各式數位資料(如:文字、圖片、PDF)並使其易於取用。

點此認識 DSpace
DSpace logo
English
中文
  • 瀏覽論文
    • 校院系所
    • 出版年
    • 作者
    • 標題
    • 關鍵字
    • 指導教授
  • 搜尋 TDR
  • 授權 Q&A
    • 我的頁面
    • 接受 E-mail 通知
    • 編輯個人資料
  1. NTU Theses and Dissertations Repository
  2. 工學院
  3. 化學工程學系
請用此 Handle URI 來引用此文件: http://tdr.lib.ntu.edu.tw/jspui/handle/123456789/90494
完整後設資料紀錄
DC 欄位值語言
dc.contributor.advisor林祥泰zh_TW
dc.contributor.advisorShiang-Tai Linen
dc.contributor.author孫肇廷zh_TW
dc.contributor.authorChao-Ting Sunen
dc.date.accessioned2023-10-03T16:20:20Z-
dc.date.available2023-11-09-
dc.date.copyright2023-10-03-
dc.date.issued2023-
dc.date.submitted2023-08-08-
dc.identifier.citationMullard, A., The drug-maker's guide to the galaxy. Nature, 2017. 549(7673): p. 445-447.
An, W.F. and N. Tolliday, Cell-Based Assays for High-Throughput Screening. Molecular Biotechnology, 2010. 45(2): p. 180-186.
Shoichet, B.K., Virtual screening of chemical libraries. Nature, 2004. 432(7019): p. 862-865.
van Hilten, N., F. Chevillard, and P. Kolb, Virtual Compound Libraries in Computer-Assisted Drug Discovery. Journal of Chemical Information and Modeling, 2019. 59(2): p. 644-651.
Alshehri, A.S., R. Gani, and F. You, Deep learning and knowledge-based methods for computer-aided molecular design—toward a unified approach: State-of-the-art and future directions. Computers & Chemical Engineering, 2020. 141: p. 107005.
Austin, N.D., N.V. Sahinidis, and D.W. Trahan, Computer-aided molecular design: An introduction and review of tools, applications, and solution techniques. Chemical Engineering Research and Design, 2016. 116: p. 2-26.
Venkatasubramanian, V., K. Chan, and J.M. Caruthers, Evolutionary Design of Molecules with Desired Properties Using the Genetic Algorithm. Journal of Chemical Information and Computer Sciences, 1995. 35(2): p. 188-195.
Kingma, D.P. and M. Welling, Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
Goodfellow, I.J., et al., Generative adversarial nets, in Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2. 2014, MIT Press: Montreal, Canada. p. 2672–2680.
Segler, M.H., et al., Generating focused molecule libraries for drug discovery with recurrent neural networks. ACS central science, 2018. 4(1): p. 120-131.
Gómez-Bombarelli, R., et al., Automatic Chemical Design Using a Data-Driven Continuous Representation of Molecules. ACS Central Science, 2018. 4(2): p. 268-276.
Kusner, M.J., B. Paige, and J.M. Hernández-Lobato. Grammar variational autoencoder. in International conference on machine learning. 2017. PMLR.
Dai, H., et al., Syntax-directed variational autoencoder for structured data. arXiv preprint arXiv:1802.08786, 2018.
Jin, W., R. Barzilay, and T. Jaakkola, Junction Tree Variational Autoencoder for Molecular Graph Generation, in Proceedings of the 35th International Conference on Machine Learning, D. Jennifer and K. Andreas, Editors. 2018, PMLR: Proceedings of Machine Learning Research. p. 2323--2332.
Liu, Q., et al., Constrained graph variational autoencoders for molecule design. Advances in neural information processing systems, 2018. 31.
Samanta, B., et al., Nevae: A deep generative model for molecular graphs. The Journal of Machine Learning Research, 2020. 21(1): p. 4556-4588.
Popova, M., O. Isayev, and A. Tropsha, Deep reinforcement learning for de novo drug design. Science Advances, 2018. 4(7): p. eaap7885.
Olivecrona, M., et al., Molecular de-novo design through deep reinforcement learning. Journal of Cheminformatics, 2017. 9(1): p. 48.
Guimaraes, G.L., et al., Objective-reinforced generative adversarial networks (organ) for sequence generation models. arXiv preprint arXiv:1705.10843, 2017.
Sanchez-Lengeling, B., et al., Optimizing distributions over molecular space. An objective-reinforced generative adversarial network for inverse-design chemistry (ORGANIC). 2017.
Putin, E., et al., Adversarial Threshold Neural Computer for Molecular de Novo Design. Molecular Pharmaceutics, 2018. 15(10): p. 4386-4397.
Lim, J., et al., Molecular generative model based on conditional variational autoencoder for de novo molecular design. Journal of cheminformatics, 2018. 10(1): p. 1-9.
Lim, J., et al., Scaffold-based molecular design with a graph generative model. Chemical Science, 2020. 11(4): p. 1153-1164.
Bagal, V., et al., Molgpt: Molecular generation using a transformer-decoder model. Journal of Chemical Information and Modeling, 2021.
Kim, H., J. Na, and W.B. Lee, Generative Chemical Transformer: Neural Machine Learning of Molecular Geometric Structures from Chemical Language via Attention. Journal of Chemical Information and Modeling, 2021. 61(12): p. 5804-5814.
Kim, S., et al., PubChem 2023 update. Nucleic Acids Research, 2022. 51(D1): p. D1373-D1380.
Ramakrishnan, R., et al., Quantum chemistry structures and properties of 134 kilo molecules. Scientific Data, 2014. 1(1): p. 140022.
Sterling, T. and J.J. Irwin, ZINC 15–ligand discovery for everyone. Journal of chemical information and modeling, 2015. 55(11): p. 2324-2337.
Gaulton, A., et al., The ChEMBL database in 2017. Nucleic Acids Research, 2016. 45(D1): p. D945-D954.
Wishart, D.S., et al., DrugBank 5.0: a major update to the DrugBank database for 2018. Nucleic acids research, 2018. 46(D1): p. D1074-D1082.
Pence, H.E. and A. Williams, ChemSpider: an online chemical information resource. 2010, ACS Publications.
Gilson, M.K., et al., BindingDB in 2015: a public database for medicinal chemistry, computational chemistry and systems pharmacology. Nucleic acids research, 2016. 44(D1): p. D1045-D1053.
Irwin, J.J., et al., ZINC: a free tool to discover chemistry for biology. Journal of chemical information and modeling, 2012. 52(7): p. 1757-1768.
Elton, D.C., et al., Deep learning for molecular design—a review of the state of the art. Molecular Systems Design & Engineering, 2019. 4(4): p. 828-849.
Gebauer, N., M. Gastegger, and K. Schütt, Symmetry-adapted generation of 3d point sets for the targeted discovery of molecules. Advances in neural information processing systems, 2019. 32.
Simm, G., R. Pinsler, and J.M. Hernández-Lobato. Reinforcement learning for molecular design guided by quantum mechanics. in International Conference on Machine Learning. 2020. PMLR.
Weininger, D., SMILES, a chemical language and information system. 1. Introduction to methodology and encoding rules. Journal of chemical information and computer sciences, 1988. 28(1): p. 31-36.
Bjerrum, E.J., SMILES enumeration as data augmentation for neural network modeling of molecules. arXiv preprint arXiv:1703.07076, 2017.
O’Boyle, N.M., Towards a Universal SMILES representation-A standard method to generate canonical SMILES based on the InChI. Journal of cheminformatics, 2012. 4: p. 1-14.
Schneider, N., R.A. Sayle, and G.A. Landrum, Get Your Atoms in Order—An Open-Source Implementation of a Novel and Robust Molecular Canonicalization Algorithm. Journal of Chemical Information and Modeling, 2015. 55(10): p. 2111-2120.
Weininger, D., A. Weininger, and J.L. Weininger, SMILES. 2. Algorithm for generation of unique SMILES notation. Journal of chemical information and computer sciences, 1989. 29(2): p. 97-101.
O'Boyle, N. and A. Dalke, DeepSMILES: an adaptation of SMILES for use in machine-learning of chemical structures. 2018.
Krenn, M., et al., SELFIES: a robust representation of semantically constrained graphs with an example application in chemistry. arXiv preprint arXiv:1905.13741, 2019.
Polykovskiy, D., et al., Molecular sets (MOSES): a benchmarking platform for molecular generation models. Frontiers in pharmacology, 2020. 11: p. 565644.
Zhu, Y., et al., A survey on deep graph generation: Methods and applications. arXiv preprint arXiv:2203.06714, 2022.
Li, Y., et al., Learning deep generative models of graphs. arXiv preprint arXiv:1803.03324, 2018.
Jin, W., R. Barzilay, and T. Jaakkola. Hierarchical generation of molecular graphs using structural motifs. in International conference on machine learning. 2020. PMLR.
He, J., et al., Molecular optimization by capturing chemist’s intuition using deep neural networks. Journal of cheminformatics, 2021. 13(1): p. 1-17.
Madhawa, K., et al., Graphnvp: An invertible flow model for generating molecular graphs. arXiv preprint arXiv:1905.11600, 2019.
Zang, C. and F. Wang, MoFlow: An Invertible Flow Model for Generating Molecular Graphs, in Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2020, Association for Computing Machinery: Virtual Event, CA, USA. p. 617–626.
Rezende, D. and S. Mohamed. Variational inference with normalizing flows. in International conference on machine learning. 2015. PMLR.
Sohl-Dickstein, J., et al. Deep unsupervised learning using nonequilibrium thermodynamics. in International Conference on Machine Learning. 2015. PMLR.
Weiss, T., et al., Guided Diffusion for Inverse Molecular Design. 2023.
Hopfield, J.J., Neural networks and physical systems with emergent collective computational abilities. Proceedings of the national academy of sciences, 1982. 79(8): p. 2554-2558.
Rumelhart, D.E., G.E. Hinton, and R.J. Williams, Learning representations by back-propagating errors. Nature, 1986. 323(6088): p. 533-536.
Van Den Oord, A., N. Kalchbrenner, and K. Kavukcuoglu. Pixel recurrent neural networks. in International conference on machine learning. 2016. PMLR.
Vaswani, A., et al., Attention is all you need. Advances in neural information processing systems, 2017. 30.
Popova, M., et al., MolecularRNN: Generating realistic molecular graphs with optimized properties. 2019.
Moret, M., et al., Generative molecular design in low data regimes. Nature Machine Intelligence, 2020. 2(3): p. 171-180.
Gupta, A., et al., Generative recurrent networks for de novo drug design. Molecular informatics, 2018. 37(1-2): p. 1700111.
Merk, D., et al., De novo design of bioactive small molecules by artificial intelligence. Molecular informatics, 2018. 37(1-2): p. 1700153.
Blaschke, T., et al., REINVENT 2.0: An AI Tool for De Novo Drug Design. Journal of Chemical Information and Modeling, 2020. 60(12): p. 5918-5922.
Langevin, M., et al., Scaffold-Constrained Molecular Generation. Journal of Chemical Information and Modeling, 2020. 60(12): p. 5637-5646.
Li, Y., et al., DeepScaffold: a comprehensive tool for scaffold-based de novo drug discovery using deep learning. Journal of chemical information and modeling, 2019. 60(1): p. 77-91.
Kotsias, P.-C., et al., Direct steering of de novo molecular generation with descriptor conditional recurrent neural networks. Nature Machine Intelligence, 2020. 2(5): p. 254-265.
Hochreiter, S. and J. Schmidhuber, Long short-term memory. Neural computation, 1997. 9(8): p. 1735-1780.
Cho, K., et al., Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078, 2014.
Luong, M.-T., H. Pham, and C.D. Manning, Effective approaches to attention-based neural machine translation. arXiv preprint arXiv:1508.04025, 2015.
Raffel, C., et al., Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 2020. 21(1): p. 5485-5551.
Lewis, M., et al., Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461, 2019.
Devlin, J., et al., Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
Radford, A., et al., Improving language understanding by generative pre-training. 2018.
Irwin, R., et al., Chemformer: a pre-trained transformer for computational chemistry. Machine Learning: Science and Technology, 2022. 3(1): p. 015022.
He, J., et al., Transformer neural network for structure constrained molecular optimization. 2021.
Yang, L., et al., Transformer-based deep learning method for optimizing ADMET properties of lead compounds. Physical Chemistry Chemical Physics, 2023.
Griffen, E., et al., Matched molecular pairs as a medicinal chemistry tool: miniperspective. Journal of medicinal chemistry, 2011. 54(22): p. 7739-7750.
Rothchild, D., et al., C5t5: Controllable generation of organic molecules with transformers. arXiv preprint arXiv:2108.10307, 2021.
Dollar, O., et al., Attention-based generative models for de novo molecular design. Chemical Science, 2021. 12(24): p. 8362-8372.
Yang, Y., et al., SyntaLinker: automatic fragment linking with deep conditional transformer neural networks. Chemical science, 2020. 11(31): p. 8312-8322.
Bemis, G.W. and M.A. Murcko, The properties of known drugs. 1. Molecular frameworks. Journal of medicinal chemistry, 1996. 39(15): p. 2887-2893.
Kramer, M.A., Nonlinear principal component analysis using autoassociative neural networks. AIChE journal, 1991. 37(2): p. 233-243.
Griffiths, R.-R. and J.M. Hernández-Lobato, Constrained Bayesian optimization for automatic chemical design using variational autoencoders. Chemical science, 2020. 11(2): p. 577-586.
Simonovsky, M. and N. Komodakis. Graphvae: Towards generation of small graphs using variational autoencoders. in Artificial Neural Networks and Machine Learning–ICANN 2018: 27th International Conference on Artificial Neural Networks, Rhodes, Greece, October 4-7, 2018, Proceedings, Part I 27. 2018. Springer.
Arjovsky, M., S. Chintala, and L. Bottou. Wasserstein generative adversarial networks. in International conference on machine learning. 2017. PMLR.
Makhzani, A., et al., Adversarial autoencoders. arXiv preprint arXiv:1511.05644, 2015.
Méndez-Lucio, O., et al., De novo generation of hit-like molecules from gene expression signatures using artificial intelligence. Nature Communications, 2020. 11(1): p. 10.
Prykhodko, O., et al., A de novo molecular generation method using latent vector based generative adversarial network. Journal of Cheminformatics, 2019. 11(1): p. 74.
Putin, E., et al., Reinforced Adversarial Neural Computer for de Novo Molecular Design. Journal of Chemical Information and Modeling, 2018. 58(6): p. 1194-1204.
You, J., et al., Graph convolutional policy network for goal-directed molecular graph generation. Advances in neural information processing systems, 2018. 31.
Maziarka, Ł., et al., Mol-CycleGAN: a generative model for molecular optimization. Journal of Cheminformatics, 2020. 12(1): p. 1-18.
De Cao, N. and T. Kipf, MolGAN: An implicit generative model for small molecular graphs. arXiv preprint arXiv:1805.11973, 2018.
Brown, N., et al., GuacaMol: Benchmarking Models for de Novo Molecular Design. Journal of Chemical Information and Modeling, 2019. 59(3): p. 1096-1108.
Bahdanau, D., K. Cho, and Y. Bengio, Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473, 2014.
Nigam, A., et al., Augmenting genetic algorithms with deep neural networks for exploring the chemical space. arXiv preprint arXiv:1909.11655, 2019.
Ahn, S., et al., Guiding deep molecular optimization with genetic exploration. Advances in neural information processing systems, 2020. 33: p. 12008-12021.
Ertl, P. and A. Schuffenhauer, Estimation of synthetic accessibility score of drug-like molecules based on molecular complexity and fragment contributions. Journal of Cheminformatics, 2009. 1(1): p. 8.
Leo, A., C. Hansch, and D. Elkins, Partition coefficients and their uses. Chemical Reviews, 1971. 71(6): p. 525-616.
Ertl, P., B. Rohde, and P. Selzer, Fast Calculation of Molecular Polar Surface Area as a Sum of Fragment-Based Contributions and Its Application to the Prediction of Drug Transport Properties. Journal of Medicinal Chemistry, 2000. 43(20): p. 3714-3717.
Bickerton, G.R., et al., Quantifying the chemical beauty of drugs. Nature Chemistry, 2012. 4(2): p. 90-98.
Testa, B., et al., Molecular drug properties: measurement and prediction. 2007: Wiley Online Library.
Ertl, P., S. Roggo, and A. Schuffenhauer, Natural Product-likeness Score and Its Application for Prioritization of Compound Libraries. Journal of Chemical Information and Modeling, 2008. 48(1): p. 68-74.
RDKit: Open-source cheminformatics. https://www.rdkit.org.
Jin, W., et al., Learning multimodal graph-to-graph translation for molecular optimization. arXiv preprint arXiv:1812.01070, 2018.
Jin, W., R. Barzilay, and T. Jaakkola, Hierarchical graph-to-graph translation for molecules. arXiv preprint arXiv:1907.11223, 2019.
Arús-Pous, J., et al., SMILES-based deep generative scaffold decorator for de-novo drug design. Journal of Cheminformatics, 2020. 12(1): p. 38.
Russell, S.J., Artificial intelligence a modern approach. 2010: Pearson Education, Inc.
Boser, B.E., I.M. Guyon, and V.N. Vapnik. A training algorithm for optimal margin classifiers. in Proceedings of the fifth annual workshop on Computational learning theory. 1992.
Cortes, C. and V. Vapnik, Support-vector networks. Machine learning, 1995. 20: p. 273-297.
Breiman, L., et al. Classification and Regression Trees. 1984.
McCulloch, W.S. and W. Pitts, A logical calculus of the ideas immanent in nervous activity. The bulletin of mathematical biophysics, 1943. 5(4): p. 115-133.
Rosenblatt, F., The perceptron: a probabilistic model for information storage and organization in the brain. Psychological review, 1958. 65(6): p. 386.
Fukushima, K., Cognitron: A self-organizing multilayered neural network. Biological Cybernetics, 1975. 20(3): p. 121-136.
Lecun, Y., et al., Gradient-based learning applied to document recognition. Proceedings of the IEEE, 1998. 86(11): p. 2278-2324.
Robbins, H. and S. Monro, A stochastic approximation method. The annals of mathematical statistics, 1951: p. 400-407.
Rombach, R., et al. High-resolution image synthesis with latent diffusion models. in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
Dosovitskiy, A., et al. CARLA: An open urban driving simulator. in Conference on robot learning. 2017. PMLR.
Bubeck, S., et al., Sparks of artificial general intelligence: Early experiments with gpt-4. arXiv preprint arXiv:2303.12712, 2023.
Tschannen, M., O. Bachem, and M. Lucic, Recent advances in autoencoder-based representation learning. arXiv preprint arXiv:1812.05069, 2018.
Wang, Y., H. Yao, and S. Zhao, Auto-encoder based dimensionality reduction. Neurocomputing, 2016. 184: p. 232-242.
Odaibo, S., Tutorial: Deriving the standard variational autoencoder (vae) loss function. arXiv preprint arXiv:1907.08956, 2019.
Higgins, I., et al. beta-vae: Learning basic visual concepts with a constrained variational framework. in International conference on learning representations. 2017.
Sohn, K., H. Lee, and X. Yan, Learning structured output representation using deep conditional generative models. Advances in neural information processing systems, 2015. 28.
Hochreiter, S., Untersuchungen zu dynamischen neuronalen Netzen. Diploma, Technische Universität München, 1991. 91(1).
Hochreiter, S., The vanishing gradient problem during learning recurrent neural nets and problem solutions. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, 1998. 6(02): p. 107-116.
Sutskever, I., O. Vinyals, and Q.V. Le, Sequence to sequence learning with neural networks. Advances in neural information processing systems, 2014. 27.
Gehring, J., et al., Convolutional Sequence to Sequence Learning, in Proceedings of the 34th International Conference on Machine Learning, P. Doina and T. Yee Whye, Editors. 2017, PMLR: Proceedings of Machine Learning Research. p. 1243--1252.
Geva, M., et al., Transformer feed-forward layers are key-value memories. arXiv preprint arXiv:2012.14913, 2020.
He, K., et al. Deep residual learning for image recognition. in Proceedings of the IEEE conference on computer vision and pattern recognition. 2016.
Ba, J.L., J.R. Kiros, and G.E. Hinton, Layer normalization. arXiv preprint arXiv:1607.06450, 2016.
Bowman, S.R., et al., Generating sentences from a continuous space. arXiv preprint arXiv:1511.06349, 2015.
Li, X. and D. Fourches, SMILES Pair Encoding: A Data-Driven Substructure Tokenization Algorithm for Deep Learning. Journal of Chemical Information and Modeling, 2021. 61(4): p. 1560-1569.
Degen, J., et al., On the Art of Compiling and Using'Drug‐Like'Chemical Fragment Spaces. ChemMedChem: Chemistry Enabling Drug Discovery, 2008. 3(10): p. 1503-1507.
Preuer, K., et al., Fréchet ChemNet Distance: A Metric for Generative Models for Molecules in Drug Discovery. Journal of Chemical Information and Modeling, 2018. 58(9): p. 1736-1741.
Loshchilov, I. and F. Hutter, Sgdr: Stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983, 2016.
Shoemake, K. Animating rotation with quaternion curves. in Proceedings of the 12th annual conference on Computer graphics and interactive techniques. 1985.
Vig, J., A multiscale visualization of attention in the transformer model. arXiv preprint arXiv:1906.05714, 2019.
-
dc.identifier.urihttp://tdr.lib.ntu.edu.tw/jspui/handle/123456789/90494-
dc.description.abstract機器學習中的深度生成模型具有快速且不受理論模型限制的優點。隨著計算能力、模型、最佳化算法和大型開源分子資料庫的演進,這些模型如今已被廣泛應用於各種分子設計任務。本研究的目的是將一種深度生成模型應用於多個分子設計任務,以實現快速且精確的生成符合目標條件的分子。
我們利用MOSES基準平台提供的158萬個中性分子,使用SMILES代表這些分子,並基於條件變分自編碼器(CVAE)訓練一個Transformer模型。它能夠在無條件、性質條件、結構條件,以及性質和結構條件的組合下生成SMILES。性質條件採用分配係數(partition coefficient)、拓撲極性表面積(topological polar surface area)和藥物相似定量估計(quantitative estimate of drug-likeness),而結構條件使用Bemis-Murcko骨架(scaffold)。
訓練後,負責不同任務的模型各自生成了大量的SMILES。我們通過有效性(validity)驗證模型對SMILES和化學結構規則的理解,通過新穎性(novelty)評估模型發現新分子的能力,並通過獨特性(uniqueness)和內部多樣性(internal diversity)驗證模型生成不同且多樣分子的能力。通過與訓練數據的比較,驗證模型是否準確學習了訓練數據的分佈。對於以性質為條件的模型,我們計算生成分子的系統和絕對誤差。對於以結構為條件的模型,我們則計算符合結構條件的比例。
結果顯示,在沒有任何約束條件下所生成的SMILES具備極高的有效性、獨特性與多樣性,並與訓練數據中的12個分子描述符(descriptor)的分佈幾乎一致。在性質約束下所生成的SMILES中接近一半符合性質條件,而且模型能夠在沒有任何訓練數據的區域中發現新的分子結構。在結構約束下,無論是模型在訓練過程中看過或沒看過的結構,幾乎所有生成的SMILES都能符合條件。甚至簡單碳鏈也能作為結構條件。在結構和性質的組合約束下,約10%生成的SMILES符合條件。不過在結構條件不大以及性質條件附近的訓練數據充足的情況下,對於未見過的結構條件生成的SMILES仍能達到一半符合條件。
另外,由於變分自編碼器能夠創造稠密且平滑(smooth)的潛在空間(latent space)作為分子表示,因此可以在潛在空間中進行內插以創造與兩分子相似結構的分子。我們通過多次的分子內插計算生成分子的結構相似度並觀察到結構的平滑性。我們量化了這種平滑性,並展示了變分自編碼器相較於自編碼器在分子內插方面的優越性。
zh_TW
dc.description.abstractDeep generative models in machine learning have the advantages of being fast and not limited by theoretical models. With the evolution of computing power, models, optimization algorithms, and large open-source molecular databases, these models are now extensively used across various tasks in molecular design. This study aims to apply a deep generative model to several molecular design tasks with the goal of rapidly and precisely creating molecules that fulfill specified conditions.
We leverage 1.58 million neutral molecules from the MOSES benchmark platform, represent these molecules using SMILES, and train a Transformer model based on a conditional variational autoencoder (CVAE). It can generate SMILES under unrestricted conditions, property-based conditions, a structural condition, and a combination of both property and structural conditions. The property conditions employ the partition coefficient, topological polar surface area, and quantitative estimate of drug-likeness, while the structural condition is defined by the Bemis-Murcko scaffold.
Following the training, the models responsible for different tasks each generate a substantial number of SMILES. We then validate the models’ understanding of the syntactic and semantic rules of SMILES through validity and assess their ability to discover new, unique, and diverse molecules through novelty, uniqueness, and internal diversity. Comparisons with the training data are used to verify whether the training data distribution has been accurately learned. For property-conditioned models, we compute the systematic and absolute errors. For structure-conditioned models, we measure the percentage of the generated molecules that adhere to the specified structure.
Results show that, without any constraints, the model can generate highly valid, novel, unique, and diverse SMILES that align with the distribution of the 12 molecular descriptors in the training data. Under property constraints, close to half of the SMILES result in molecules conforming to property conditions, and the model can discover novel molecular structures in areas without any training data. Under a structural constraint, the model can nearly perfectly generate molecules that conform to seen and unseen structures and can be constrained by simple carbon chains that are not Bemis-Murcko scaffolds. Under combined structural and property constraints, about 10% of the generated SMILES conform to the conditions. However, in cases where the structural constraint is minimal and there is ample training data near the property condition, about half of the SMILES generated under unseen structural conditions still meet the criteria
Moreover, due to the variational autoencoder's ability to create a dense and smooth latent space for representing molecules, molecules with structures similar to two given molecules can be created through interpolation in this latent space. We calculate the structural similarity of the generated molecules through multiple molecular interpolations and observe the smoothness of the structures. We quantify this smoothness and demonstrate the superiority of variational autoencoders over standard autoencoders in terms of molecular interpolation.
en
dc.description.provenanceSubmitted by admin ntu (admin@lib.ntu.edu.tw) on 2023-10-03T16:20:20Z
No. of bitstreams: 0
en
dc.description.provenanceMade available in DSpace on 2023-10-03T16:20:20Z (GMT). No. of bitstreams: 0en
dc.description.tableofcontents誌謝 i
中文摘要 ii
Abstract iii
Table of Contents iv
List of Figures viii
List of Tables xiii
Chapter 1 Introduction 1
1.1 Exploring the Chemical Space 1
1.2 Molecular Representation 4
1.2.1 One-Dimensional Representation 4
1.2.2 Two-Dimensional Representation 5
1.2.3 Comparison of Two Representations 6
1.3 DGMs for Molecular Design 6
1.3.1 Autoregressive Models 7
1.3.2 Variational Autoencoders 9
1.3.3 Generative Adversarial Nets 11
1.4 Tasks of Molecular Design 12
1.4.1 Unconstrained Generation 13
1.4.2 Property-Constrained Generation 15
1.4.3 Structure-Constrained Generation 17
1.5 Our Model for Multi-Task Molecular Design 20
Chapter 2 Theory 23
2.1 Machine Learning 23
2.1.1 Artificial Intelligence and Machine Learning 23
2.1.2 Three Learning Paradigms 23
2.2 Deep Learning 25
2.2.1 Deep Learning over Classical Machine Learning 25
2.2.2 Deep Learning Architecture 26
2.2.3 Workflow of Deep Learning 27
2.3 Deep Generative Models 28
2.3.1 Autoencoder 29
2.3.2 Variational Autoencoder 30
2.4 Evolving from RNN to Self-Attention 36
2.4.1 Word Embedding 36
2.4.2 Recurrent Neural Network 38
2.4.3 Recurrent Neural Network with Attention Mechanism 40
2.4.4 Transformer 42
Chapter 3 Computational Details 49
3.1 Model Architecture 49
3.2 Hyper-Parameters 51
3.2.1 Overall Hyper-Parameters 51
3.2.2 Learning Rate Scheduling 52
3.2.3 Kullback-Leibler Annealing 53
3.3 Dataset 54
3.3.1 Property Analysis and Preprocessing 55
3.3.2 Scaffold Distribution 58
3.3.3 SMILES Preprocessing 60
3.4 SMILES Enumeration 62
3.5 Decoding Algorithm 62
3.6 Model Selection 65
3.7 Unconditioned and Conditioned Generation 69
3.7.1 Unconditioned Generation 69
3.7.2 Property-Conditioned Generation 71
3.7.3 Structure-Conditioned Generation 72
3.7.4 Property and Scaffold Conditioned Generation 74
3.8 Training and Inference Time 74
3.9 Molecular Interpolation 75
Chapter 4 Results and Discussions 78
4.1 Exploring Chemical Space Distributions 79
4.1.1 Benchmarking on MOSES 79
4.1.2 Distribution Plots for Molecular Descriptors 83
4.1.3 Exploring Novel Bemis-Murcko Scaffolds 86
4.2 Generating Molecules with Target Properties 88
4.2.1 Metrics of Basic Requirements 88
4.2.2 Property Distributions and Errors 89
4.2.3 Case Study: logP, tPSA, QED = (1, 30, 0.6) 93
4.3 Generating Molecules with Target Scaffolds 95
4.3.1 Metrics of Basic Requirements 95
4.3.2 Case Study: Small and Large Scaffold Conditions 99
4.3.3 Case Study: A Non-Scaffold as the Scaffold Condition 101
4.3.4 Comparing SCA-VAETF with MolGPT 104
4.4 Generating Molecules with Target Properties and a Scaffold 105
4.4.1 Metrics of Basic Requirements 105
4.4.2 Property Distributions and Errors 109
4.4.3 Two Major Limitations 115
4.4.4 Case Study: The Best and Worst Cases 117
4.4.5 Comparing PSCA-VAETF with MolGPT 120
4.5 Summary of The Four Tasks 122
4.6 Interpolating Molecules via Latent Space 125
4.7 Improving by Enumerating SMILES 136
4.8 Attention Visualization 139
Chapter 5 Conclusions and Future Prospects 142
REFERENCE 145
-
dc.language.isoen-
dc.subject機器學習zh_TW
dc.subject條件變分自編碼器zh_TW
dc.subject深度生成模型zh_TW
dc.subjectTransformer模型zh_TW
dc.subject多任務分子設計zh_TW
dc.subjectdeep generative modelen
dc.subjectconditional variational autoencoderen
dc.subjectTransformeren
dc.subjectmachine learningen
dc.subjectmulti-task molecular designen
dc.title運用條件式變分自編碼器進行多任務分子設計zh_TW
dc.titleMulti-Task Molecular Design Using Conditional Variational Autoencoder Based Transformeren
dc.typeThesis-
dc.date.schoolyear111-2-
dc.description.degree碩士-
dc.contributor.oralexamcommittee余柏毅;謝介銘;李奕霈zh_TW
dc.contributor.oralexamcommitteeBor-Yih Yu;Chieh-Ming Hsieh;Yi-Pei Lien
dc.subject.keyword機器學習,深度生成模型,條件變分自編碼器,Transformer模型,多任務分子設計,zh_TW
dc.subject.keywordmachine learning,deep generative model,conditional variational autoencoder,Transformer,multi-task molecular design,en
dc.relation.page151-
dc.identifier.doi10.6342/NTU202302721-
dc.rights.note同意授權(限校園內公開)-
dc.date.accepted2023-08-09-
dc.contributor.author-college工學院-
dc.contributor.author-dept化學工程學系-
dc.date.embargo-lift2025-12-31-
顯示於系所單位:化學工程學系

文件中的檔案:
檔案 大小格式 
ntu-111-2.pdf
授權僅限NTU校內IP使用(校園外請利用VPN校外連線服務)
13.91 MBAdobe PDF
顯示文件簡單紀錄


系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。

社群連結
聯絡資訊
10617臺北市大安區羅斯福路四段1號
No.1 Sec.4, Roosevelt Rd., Taipei, Taiwan, R.O.C. 106
Tel: (02)33662353
Email: ntuetds@ntu.edu.tw
意見箱
相關連結
館藏目錄
國內圖書館整合查詢 MetaCat
臺大學術典藏 NTU Scholars
臺大圖書館數位典藏館
本站聲明
© NTU Library All Rights Reserved