RetroXpert: Decompose Retrosynthesis Prediction Like A Chemist

Authors: Chaochao Yan, Qianggang Ding, Peilin Zhao, Shuangjia Zheng, JINYU YANG, Yang Yu, Junzhou Huang

NeurIPS 2020 | Venue PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We evaluate our method on USPTO-50K [19] and USPTO-full [25] to verify its effectiveness and scalability. ...Our method Retro Xpert achieves impressive performance on the test data. ...Experimental results are reported at the bottom of Table 3.
Researcher Affiliation Collaboration Chaochao Yan University of Texas at Arlington EMAIL Qianggang Ding Tsinghua University EMAIL Peilin Zhao Tencent AI Lab EMAIL Shuangjia Zheng Sun Yat-sen University EMAIL Jinyu Yang University of Texas at Arlington EMAIL Yang Yu Tencent AI Lab EMAIL Junzhou Huang University of Texas at Arlington EMAIL
Pseudocode No The paper describes the architecture and functionality of its models (EGAT, RGN) using mathematical equations and textual explanations. However, it does not include any explicitly labeled pseudocode blocks or algorithms in a structured format.
Open Source Code Yes Code and processed USPTO-full data are available at https://github.com/uta-smile/Retro Xpert
Open Datasets Yes We evaluate our method on USPTO-50K [19] and USPTO-full [25] to verify its effectiveness and scalability.
Dataset Splits Yes We adopt the same training/validation/test splits in 8:1:1 as [12, 5].
Hardware Specification Yes We train the RGN for 300, 000 time steps, and it takes about 30 hours on two GTX 1080 Ti GPUs.
Software Dependencies No The paper mentions software used such as 'DGL [30]', 'Open NMT [33]', and 'RDKit 4'. However, it does not specify version numbers for these software components, which is required for a reproducible description of ancillary software.
Experiment Setup Yes As for the EGAT, we stack three identical four-head attentive layers of which the hidden dimension is 128. All embedding sizes in EGAT are set to 128, such as F, F , and D. The Nmax is set to be two to cover 99.97% training samples. We train the EGAT on USPTO-50K for 80 epochs. EGAT parameters are optimized with Adam [34] with default settings, and the initial learning rate is 0.0005 and it is scheduled to multiply 0.2 every 20 epochs.