2306.17702v1
2306.17702v1
2306.17702v1
task in the drug discovery pipeline, which has re- computed or handcrafted molecular fingerprints; Sequence-
cently gained considerable attention thanks to ad- based neural architectures including Recurrent Neural Net-
vances in deep neural networks. However, recent works (RNNs) (Medsker & Jain, 1999), 1D Convolutional
research has revealed that deep models struggle Neural Networks (1D CNNs) (Gu et al., 2018), and Trans-
to beat traditional non-deep ones on MPP. In this formers (Honda et al., 2019; Rong et al., 2020) are exploited
study, we benchmark 12 representative models to encode molecules represented in Simplified Molecular-
(3 non-deep models and 9 deep models) on 14 Input Line-Entry System (SMILES) strings (Weininger
molecule datasets. Through the most comprehen- et al., 1989). Later, it is argued that molecules can be natu-
sive study to date, we make the following key ob- rally represented in graph structures with atoms as nodes and
servations: (i) Deep models are generally unable bonds as edges. This inspires a line of works to leverage
to outperform non-deep ones; (ii) The failure of such structured inductive bias for better molecular repre-
deep models on MPP cannot be solely attributed sentations (Gilmer et al., 2017; Xiong et al., 2019; Yang
to the small size of molecular datasets. What mat- et al., 2019; Song et al., 2020). The key advancements
ters is the irregular molecule data pattern; (iii) underneath these approaches are Graph Neural Networks
In particular, tree models using molecular finger- (GNNs), which consider graph structures and attributive
prints as inputs tend to perform better than other features simultaneously by recursively aggregating node
competitors. Furthermore, we conduct extensive features from neighborhoods (Kipf & Welling, 2017; Velick-
empirical investigations into the unique patterns ovic et al., 2018; Hamilton et al., 2017). More recently,
of molecule data and inductive biases of various researchers incorporate 3D conformations of molecules into
models underlying these phenomena. their representations for better performance, whereas prag-
matic considerations such as calculation cost, alignment
invariance, and uncertainty in conformation generation lim-
1. Introduction ited the practical applicability of these models (Axen et al.,
Molecular Property Prediction (MPP) is a critical task in 2017; Gasteiger et al., 2020; Schuett et al., 2017; Gasteiger
drug discovery, aimed at identifying molecules with desir- et al., 2021; Liu et al., 2022). We summarize the widely-
able pharmacological and ADMET (absorption, distribu- used molecular descriptors and their corresponding mod-
tion, metabolism, excretion, and toxicity) properties. Ma- els in our benchmark, as shown in Figure 1. Despite the
chine learning models have been widely used in this fast- fruitful progress, previous studies (Mayr et al., 2018; Yang
growing field, with two types of models being commonly et al., 2019; Valsecchi et al., 2022; Jiang et al., 2021; van
employed: traditional non-deep models and deep models. Tilborg et al., 2022; Janela & Bajorath, 2022) have observed
In non-deep models, molecules are fed into traditional ma- that deep models struggled to outperform non-deep ones
chine learning models such as Random Forest and Support on molecules. However, these studies neither consider the
Vector Machine in the format of computed or handcrafted emerging powerful deep models (e.g., Transformer (Honda
molecular fingerprints (Todeschini & Consonni, 2010). The et al., 2019), SphereNet (Liu et al., 2021)) nor explore vari-
other group utilizes deep models to extract expressive repre- ous molecular descriptors (e.g., 3D molecular graph). Also,
they did not investigate the reasons why deep models often
*
Equal contribution 1 Westlake University, Hangzhou, China. fail on molecules.
Correspondence to: Jun Xia <[email protected]>, Stan Z.
Li <[email protected]>. To narrow this gap, we present the most comprehensive
benchmark study on molecular property prediction to date,
The ICML 2023 Workshop on Interpretable Machine Learning in with a precise methodology for dataset inclusion and hy-
Healthcare (IMLH), Honolulu, Hawaii, USA. PMLR 202, 2023.
Copyright 2023 by the author(s). perparameter tuning. Our empirical results confirm the
1
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
Atoms features
7 0 1 0 0
Descriptors Nc1ccc(cc1)C(=O)O
1 0 1 0 0 0 1 0 0 1 0 3
Bonds features
Models SVM RF XGB MLP CNN RNN TRSF GCN MPNN GAT AFP SPN
(a) Fingerprints (FPs) (b) SMILES string (c) 2D Graph (d) 3D Graph
Figure 1. Exemplary molecular descriptors and their corresponding models in our benchmark. SVM: Support Vector Machine (Zernov
et al., 2003); RF: Random Forest (Svetnik et al., 2003); XGB: eXtreme Gradient Boosting (Chen & Guestrin, 2016); MLP: Multi-Layer
Perceptron; CNN: 1D Convolution Neural Network (Kimber et al., 2021); RNN: Recurrent Neural Network (GRU) (Mulder et al., 2015);
TRSF: TRanSFormer (Vaswani et al., 2017); GCN: Graph Convolution Network (Kipf & Welling, 2017); MPNN: Message-Passing
Neural Network (Gilmer et al., 2017); GAT: Graph Attention neTwork (Velickovic et al., 2018); AFP: Attentive FP (Xiong et al., 2020);
SPN: SPhereNet (Liu et al., 2022). The above-mentioned abbreviations are applicable throughout the entire paper.
observations of previous studies, namely that deep models matters is the irregular molecule data pattern, not solely the
generally cannot outperform traditional non-deep counter- dataset size. We will provide an in-depth analysis to the
parts. Moreover, we observe several interesting phenom- unique molecule data pattern in Sec. 3.
ena that challenge the prevailing beliefs of the community,
Observation 3. Tree models (XGB and RF) exhibit a par-
which can guide optimal methodology design for future
ticular advantage over other models.
studies. Furthermore, we transform the original molecular
In the experiments shown in Table 1, we can see that the
data to observe the performance changes of various models,
tree-based models consistently rank among the top three
uncovering the unique patterns of molecular data and the
on each dataset. Additionally, tree models rank as the top
differing inductive biases of various models. These in-depth
one on 8/15 datasets. We will explore why tree models are
empirical studies shed light on the benchmarking results.
well-suited for molecular fingerprints in Sec. 3.
2
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
Table 1. The comparison of representative models on multiple molecular datasets. The standard deviations can be seen in the appendix for
the limited space. No.: Number of the molecules in the datasets. The top-3 performances on each dataset are highlighted with the grey
background. The best performance is highlighted with bold. Kindly note that ‘TRSF’ denotes the transformer that has been pre-trained
on 861, 000 molecular SMILES strings. The results on QM 9 can be seen in the appendix.
Dataset (No.) Metric SVM XGB RF CNN RNN TRSF MLP GCN MPNN GAT AFP SPN
BACE (1,513) AUC ROC 0.886 0.896 0.890 0.815 0.559 0.835 0.887 0.880 0.846 0.886 0.879 0.882
HIV (40,748) AUC ROC 0.817 0.823 0.826 0.733 0.639 0.748 0.791 0.834 0.814 0.812 0.819 0.818
BBBP (2,035) AUC ROC 0.913 0.926 0.923 0.760 0.693 0.897 0.918 0.915 0.872 0.902 0.893 0.905
ClinTox (1,475) AUC ROC 0.879 0.919 0.933 0.685 0.813 0.963 0.890 0.889 0.868 0.891 0.907 0.912
SIDER (1,366) AUC ROC 0.626 0.638 0.644 0.591 0.515 0.641 0.617 0.633 0.603 0.614 0.620 0.613
Tox21 (7,811) AUC ROC 0.820 0.837 0.838 0.766 0.734 0.817 0.834 0.830 0.816 0.829 0.845 0.827
ToxCast (8,539) AUC ROC 0.725 0.785 0.778 0.735 0.74 0.780 0.781 0.767 0.736 0.768 0.788 0.772
MUV (93,087) AUC PRC 0.093 0.072 0.069 0.045 0.094 0.059 0.018 0.056 0.019 0.055 0.044 0.058
SARS-CoV-2 (14,332) AUC ROC 0.599 0.700 0.686 0.688 0.649 0.643 0.638 0.646 0.640 0.683 0.651 0.663
ESOL (1,127) RMSE 0.676 0.583 0.647 2.569 1.511 0.718 0.653 0.773 0.695 0.661 0.594 0.671
Lipop (4,200) RMSE 0.683 0.585 0.626 1.016 1.207 0.947 0.633 0.665 0.669 0.680 0.664 0.630
FreeSolv (639) RMSE 1.063 0.715 1.014 2.275 2.205 1.504 1.046 1.316 1.327 1.304 1.139 1.159
QM7 (6,830) MAE 42.814 52.726 51.403 81.165 158.160 64.363 86.060 64.530 107.013 78.217 59.973 55.727
QM8 (21,786) MAE 0.0364 0.0126 0.0098 0.0205 0.0295 0.0232 0.0104 0.0154 0.0109 0.0187 0.0098 0.0103
RMSE
0.5 MAE
0.6 60
0.4
0.5 0.3 40
0.2
20
0.4 0.1
0-smooth 10-smooth 20-smooth 0-smooth 10-smooth 20-smooth 0-smooth 10-smooth 20-smooth
Figure 2. The performance of various models on the smoothed datasets. Left: ESOL (Regression); Middle: Lipop (Regression); Right:
QM7 (Regression). We only smooth the regression datasets because the labels of classification datasets are not suitable for smoothing.
the smoothed label ybi to train the models. The results are
shown in Figure 2, where ‘0-smooth’ denotes the original
datasets. ‘10-smooth’ and ‘20-smooth’ mean k = 10 and
k = 20, respectively. As can be observed, the performance
of deep models improves dramatically as the level of dataset
smoothing increases, and many deep models including MLP,
GCN, and AFP can even beat non-deep ones after smooth-
Figure 3. Examplary of Activity Cliffs (ACs) on the target named
ing. These phenomena indicate that deep models are more dopamine D3 receptor (D3R). Ki means the bioactivity values.
suitable for the smoothed datasets. This figure is adapted from Derek van Tilborg’s work (van Tilborg
Secondly, we decrease the level of data smoothing using et al., 2022).
the concept of activity cliff (Maggiora, 2006; Stumpfe &
Bajorath, 2012) from chemistry, which means a situation
where small changes in the chemical structure of a drug lead Tilborg et al., 2022). The test set contains molecules that
to significant changes in its bioactivity. We provide an ex- are chemically similar to those in the training set but exhibit
ample activity cliff pairs in Figure 3. Apparently, the target either a large difference in bioactivity (cliff molecules) or
function of activity cliffs that map molecules to the activity similar bioactivity (non-cliff molecules). As shown in Ta-
values is less smooth than normal molecular datasets. We ble 2, the non-deep models consistently outperform deep
then evaluate the models on the activity cliff datasets (van ones on these activity cliff datasets. Furthermore, it is worth
noting that deep models exhibit a similar level of perfor-
3
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
Table 2. RMSEnc and RMSEc are the prediction RMSE on non-cliff molecules and cliff molecules, respectively. ∆R = (RMSEc -
RMSEnc ) / RMSEnc ×100%. The top-3 performances and the best performance are highlighted with grey background and bold.
Target name
Metric SVM XGB RF CNN RNN TRSF MLP GCN MPNN GAT AFP
(Response type)
RMSEnc 0.652 0.623 0.619 0.934 0.712 0.785 0.707 0.932 0.938 0.960 0.909
CB1
RMSEc 0.773 0.767 0.770 0.944 0.823 0.888 0.807 0.992 0.989 0.975 0.967
(Agonism EC50 )
∆R 18.55% 23.11% 24.39% 1.15% 15.59% 13.12% 14.1% 6.37% 5.47% 1.55% 6.35%
RMSEnc 0.589 0.579 0.577 0.871 0.692 0.801 0.664 0.927 0.820 0.995 0.865
DAT
RMSEc 0.744 0.696 0.730 0.894 0.783 0.934 0.792 1.003 0.921 1.042 0.995
(Inhibition Ki )
∆R 26.30% 20.18% 26.64% 2.48% 13.15% 16.70% 19.40% 8.23% 12.38% 4.74% 15.11%
RMSEnc 0.535 0.552 0.561 0.854 0.696 0.799 0.606 0.856 0.833 0.892 0.749
PPARα
RMSEc 0.671 0.678 0.685 0.962 0.825 0.968 0.713 0.870 0.872 0.929 0.823
(Agonism EC50 )
∆R 25.42% 22.83% 22.10% 12.69% 15.64% 21.26% 17.77% 1.72% 4.78% 4.21% 9.90%
RMSEnc 0.598 0.592 0.591 0.938 0.893 0.873 0.663 1.095 0.958 1.102 1.018
DOR
RMSEc 0.861 0.854 0.836 1.098 1.036 1.032 0.874 1.259 1.152 1.281 1.179
(Inhibition Ki )
∆R 43.98% 44.14% 41.46% 17.06% 16.01 % 18.26% 31.85% 14.93% 20.27% 16.26% 15.83%
mance on both non-cliff and cliff molecules, while non-deep the molecular graphs, i.e., we apply orthogonal transforma-
models experience significant changes in performance when tions to both the atom features and bond features. As can
transitioning from non-cliff to cliff molecules. These phe- be observed in Figure 4, the performance of tree models
nomena indicate that deep models are less sensitive to subtle deteriorates dramatically and falls behind most deep models
structural changes and struggle to learn non-smooth target after the orthogonal transformation. It is because each di-
functions compared with tree models, especially the activity mension of xbi is a convex combination of all the dimensions
cliff cases. Our explanation is consistent with the conclu- of xi according to the matrix-vector product rule. In other
sions in deep learning theory (Rahaman et al., 2019), i.e., words, the molecular features after orthogonal transforma-
deep models struggle to learn high-frequency components tion no longer carry meanings individually, accounting for
of the target functions. However, tree models can learn the failure of tree models that make decisions based on each
piece-wise target functions, and do not exhibit such bias. dimension of the features separately. The learning style of
Our explorations uncover several promising avenues to en- tree models is more suitable for molecular data because only
hance deep models’ performance on molecules: smoothing a handful of features (e.g., certain substructures) are most
the target functions or improving deep models’ ability to indicative of molecular properties. On the other hand, the
learn the non-smooth target functions. performance decreases of deep models are less significant,
and most deep models can beat tree models after the trans-
Explanation 2. Deep models mix different dimensions of
formations. We explain this observation as follows. Without
molecular features, whereas tree models make decisions
the loss of generality, we assume that a linear layer of deep
based on each dimension of the features separately.
models can map the original molecular feature xi to the
Typically, features in molecular data carry meanings indi-
label yi ,
vidually. Each dimension of molecular fingerprints often
yi = W ⊤ xi + b, (3)
indicates whether a certain substructure is present in the
molecule; each dimension of nodes/edges features in molec- where W and b denote the parameter matrix and the bias
ular graph data indicates a specific characteristic of the term of the linear layer, respectively. And then, we aim to
atoms/bonds (e.g., atom/bond type, atom degree). To verify learn a new linear layer mapping the transformed model
the above explanation, we mix the different dimensions of feature xbi to label yi ,
molecular features xi ∈ Rd using an orthogonal transforma-
tion before feeding them into various models, c ⊤ xbi + b = W
yi = W c ⊤ Qxi + b̂, (4)
xbi = Qxi , (2) where W c and b̂ denote the parameter matrix and the bias
term of the new linear layer, respectively. Apparently, to
where Q ∈ Rd×d is the orthogonal matrix and xbi is the achieve the same results as the original feature, we only
molecular feature after transformation. Kindly note that the have to learn W c so that W c = QW because Q−1 = Q⊤
meaning of xi depends on the input molecular descriptors as an orthogonal matrix, and also b̂ = b. Therefore, ap-
in the experiments. Specifically, for SVM, XGB, RF, and plying the orthogonal transformation to molecular features
MLP, xi denotes the molecular fingerprints; for GNN mod- barely impacts the performance of deep models. The em-
els, xi can denote the atom features and bond features in pirical results in Figure 4 confirm this point although some
4
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
performance changes are observable due to uncontrollable Gilmer, J., Schoenholz, S. S., Riley, P. F., Vinyals, O.,
random factors. This explanation inspires us not to mix the and Dahl, G. E. Neural message passing for quan-
molecular features before feeding them into models. tum chemistry. In Proceedings of the 34th International
Conference on Machine Learning, pp. 1263–1272, 2017.
References Goh, G. B., Hodas, N. O., and Vishnu, A. Deep learning
Atz, K., Grisoni, F., et al. Geometric deep learning on for computational chemistry. Journal of computational
molecular representations. Nature Machine Intelligence, chemistry, 38(16):1291–1307, 2017.
2021. Gu, J., Wang, Z., Kuen, J., Ma, L., Shahroudy, A., Shuai,
Axen, S. D., Huang, X.-P., Cáceres, E. L., Gendelev, L., B., Liu, T., Wang, X., Wang, G., Cai, J., et al. Re-
Roth, B. L., and Keiser, M. J. A simple representa- cent advances in convolutional neural networks. Pattern
tion of three-dimensional molecular structure. Journal of recognition, 77:354–377, 2018.
medicinal chemistry, 60(17):7393–7409, 2017. Hamilton, W., Ying, Z., and Leskovec, J. Inductive repre-
Chen, T. and Guestrin, C. Xgboost: A scalable tree boost- sentation learning on large graphs. Advances in neural
ing system. In Proceedings of the 22nd acm sigkdd information processing systems, 30, 2017.
international conference on knowledge discovery and Honda, S., Shi, S., and Ueda, H. R. Smiles transformer: Pre-
data mining, pp. 785–794, 2016. trained molecular fingerprint for low data drug discovery.
Cho, K., Van Merriënboer, B., Gulcehre, C., Bahdanau, arXiv preprint arXiv:1911.04738, 2019.
D., Bougares, F., Schwenk, H., and Bengio, Y. Learn- Hu, B., Xia, J., Zheng, J., Tan, C., Huang, Y., Xu, Y.,
ing phrase representations using rnn encoder-decoder and Li, S. Z. Protein language models and structure
for statistical machine translation. arXiv preprint prediction: Connection and progression. arXiv preprint
arXiv:1406.1078, 2014. arXiv:2211.16742, 2022.
Coley, C. W., Barzilay, R., et al. Convolutional embedding Hu, W., Liu, B., and others. Strategies for pre-training graph
of attributed molecular graphs for physical property pre- neural networks. ICLR, 2020.
diction. Journal of chemical information and modeling,
2017. Janela, T. and Bajorath, J. Simple nearest-neighbour anal-
ysis meets the accuracy of compound potency predic-
Devlin, J., Chang, M., et al. BERT: Pre-training of Deep tions using complex machine learning models. Nature
Bidirectional Transformers for Language Understanding. Machine Intelligence, pp. 1–10, 2022.
In NAACL, 2019.
Jiang, D., Wu, Z., Hsieh, C.-Y., Chen, G., Liao, B., Wang,
Du, W., Zhang, H., et al. SE(3) Equivariant Graph Neural Z., Shen, C., Cao, D., Wu, J., and Hou, T. Could graph
Networks with Complete Local Frames. In ICML, 2022. neural networks learn better molecular representation for
drug discovery? a comparison study of descriptor-based
Duvenaud, D. K., Maclaurin, D., Iparraguirre, J., Bom-
and graph-based models. Journal of cheminformatics, 13
barell, R., Hirzel, T., Aspuru-Guzik, A., and Adams, R. P.
(1):1–23, 2021.
Convolutional networks on graphs for learning molecular
fingerprints. Advances in neural information processing Kearnes, S., McCloskey, K., et al. Molecular
systems, 28, 2015. Graph Convolutions: Moving Beyond Fingerprints.
J. Comput. Aided Mol. Des., 2016.
Gao, Z., Tan, C., Wu, L., and Li, S. Z. Cosp: Co-
supervised pretraining of pocket and ligand. arXiv Kimber, T. B., Gagnebin, M., and Volkamer, A. Maxsmi:
preprint arXiv:2206.12241, 2022. maximizing molecular property prediction performance
with confidence estimation using smiles augmentation
Gasteiger, J., Groß, J., and Günnemann, S. Direc-
and deep learning. Artificial Intelligence in the Life
tional message passing for molecular graphs. In
Sciences, 1:100014, 2021.
International Conference on Learning Representations,
2020. URL https://openreview.net/forum? Kipf, N. T. and Welling, M. Semi-Supervised Classification
id=B1eWbxStPH. with Graph Convolutional Networks. In ICLR, 2017.
Gasteiger, J., Becker, F., and Günnemann, S. Gemnet: Uni- Krenn, M., Häse, F., Nigam, A., Friederich, P., and Aspuru-
versal directional graph neural networks for molecules. Guzik, A. Self-referencing embedded strings (selfies): A
Advances in Neural Information Processing Systems, 34: 100% robust molecular string representation. Machine
6790–6802, 2021. Learning: Science and Technology, 1(4):045024, 2020.
5
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
Landrum, G. RDKit: A software suite for cheminformatics, Rahaman, N., Baratin, A., Arpit, D., Draxler, F., Lin, M.,
computational chemistry, and predictive modeling, 2013. Hamprecht, F., Bengio, Y., and Courville, A. On the spec-
tral bias of neural networks. In International Conference
Li, P., Wang, J., and others. An effective self-supervised on Machine Learning, pp. 5301–5310. PMLR, 2019.
framework for learning expressive molecular global rep-
resentations to drug discovery. BIB, 2021. Rogers, D. and Hahn, M. Extended-connectivity finger-
prints. J chem inf, 2010.
Liu, Y., Wang, L., et al. Spherical message passing for 3d
molecular graphs. In Iclr, 2021.
Rong, Y., Bian, Y., Xu, T., Xie, W., Wei, Y., Huang, W., and
Liu, Y., Wang, L., Liu, M., Lin, Y., Zhang, X., Oztekin, Huang, J. Self-supervised graph transformer on large-
B., and Ji, S. Spherical message passing for 3d molec- scale molecular data. Advances in Neural Information
ular graphs. In International Conference on Learning Processing Systems, 33:12559–12571, 2020.
Representations (ICLR), 2022.
Ross, J., Belgodere, B., et al. Molformer: Large Scale
Liu, Y., Liang, K., Xia, J., Zhou, S., Yang, X., Liu, X., and Chemical Language Representations Capture Molecular
Li, S. Z. Dink-net: Neural clustering on large graphs. Structure and Properties. Nat. Mach. Intell., 2022.
arXiv preprint arXiv:2305.18405, 2023.
Satorras, V. G., Hoogeboom, E., et al. E(n) Equivariant
Lu, C., Liu, Q., Wang, C., Huang, Z., Lin, P., and He, L. Graph Neural Networks. In ICML, 2021.
Molecular property prediction: A multilevel quantum
interactions modeling perspective. In Proceedings of the Schuett, T. K., Kindermans, and others. Schnet: A
AAAI Conference on Artificial Intelligence, volume 33, continuous-filter convolutional neural network for model-
pp. 1052–1060, 2019. ing quantum interactions. NIPS, 2017.
Maggiora, G. M. On outliers and activity cliffs why qsar Song, Y., Zheng, S., Niu, Z., Fu, Z.-H., Lu, Y., and Yang,
often disappoints, 2006. Y. Communicative representation learning on attributed
Mayr, A., Klambauer, G., Unterthiner, T., Steijaert, M., Weg- molecular graphs. In IJCAI, volume 2020, pp. 2831–
ner, J. K., Ceulemans, H., Clevert, D.-A., and Hochreiter, 2838, 2020.
S. Large-scale comparison of machine learning methods
Stumpfe, D. and Bajorath, J. Exploring activity cliffs
for drug target prediction on chembl. Chemical science,
in medicinal chemistry: miniperspective. Journal of
9(24):5441–5451, 2018.
medicinal chemistry, 55(7):2932–2942, 2012.
Medsker, L. and Jain, L. C. Recurrent neural networks:
design and applications. CRC press, 1999. Svetnik, V., Liaw, A., Tong, C., Culberson, J. C., Sheridan,
R. P., and Feuston, B. P. Random forest: a classification
Min, E., Chen, R., Bian, Y., Xu, T., Zhao, K., Huang, W., and regression tool for compound classification and qsar
Zhao, P., Huang, J., Ananiadou, S., and Rong, Y. Trans- modeling. Journal of chemical information and computer
former for graphs: An overview from architecture per- sciences, 43(6):1947–1958, 2003.
spective. arXiv preprint arXiv:2202.08455, 2022.
Tan, C., Xia, J., Wu, L., and Li, S. Z. Co-learning: Learning
Mulder, W. D., Bethard, S., and Moens, M.-F. A survey on from noisy labels with self-supervision. In Proceedings of
the application of recurrent neural networks to statistical the 29th ACM International Conference on Multimedia,
language modeling. Comput. Speech Lang., 30:61–98, pp. 1405–1413, 2021.
2015.
Tan, C., Gao, Z., Xia, J., Hu, B., and Li, S. Z. Global-
Ozaki, Y., Tanigaki, Y., Watanabe, S., and Onishi, M. Mul-
context aware generative protein design. In ICASSP
tiobjective tree-structured parzen estimator for computa-
2023-2023 IEEE International Conference on Acoustics,
tionally expensive optimization problems. Proceedings
Speech and Signal Processing (ICASSP), pp. 1–5. IEEE,
of the 2020 Genetic and Evolutionary Computation
2023.
Conference, 2020.
Ozturk, H., Ozgur, A., and Ozkirimli, E. Deepdta: deep Tian, H., Ketkar, R., and Tao, P. Admetboost: a web server
drug–target binding affinity prediction. Bioinformatics, for accurate admet prediction. Journal of Molecular
34(17):i821–i829, 2018. Modeling, 28(12):1–6, 2022.
Pattanaik, L. and Coley, C. W. Molecular representation: Todeschini, R. and Consonni, V. Molecular descriptors.
going long on fingerprints. Chem, 6(6):1204–1207, 2020. Recent Advances in QSAR Studies, pp. 29–102, 2010.
6
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
Valsecchi, C., Collarile, M., Grisoni, F., Todeschini, R., Xia, J., Zhu, Y., Du, Y., and Li, S. Z. A survey of pretraining
Ballabio, D., and Consonni, V. Predicting molecular on graphs: Taxonomy, methods, and applications. arXiv
activity on nuclear receptors by multitask neural networks. preprint arXiv:2202.07893, 2022e.
Journal of Chemometrics, 36(2):e3325, 2022.
Xia, J., Zhao, C., et al. Mole-BERT: Rethinking Pre-training
van Tilborg, D., Alenicheva, A., and Grisoni, F. Exposing Graph Neural Networks for Molecules. In ICLR, 2023a.
the limitations of molecular machine learning with activ-
Xia, J., Zhu, Y., Du, Y., Liu, Y., and Li, S. Z. A systematic
ity cliffs. Journal of Chemical Information and Modeling,
survey of chemical pre-trained models. IJCAI, 2023b.
62(23):5938–5951, 2022.
Xiong, Z., Wang, D., Liu, X., Zhong, F., Wan, X., Li, X.,
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, Li, Z., Luo, X., Chen, K., Jiang, H., et al. Pushing the
L., Gomez, A. N., Kaiser, Ł., and Polosukhin, I. At- boundaries of molecular representation for drug discovery
tention is all you need. Advances in neural information with the graph attention mechanism. Journal of medicinal
processing systems, 30, 2017. chemistry, 63(16):8749–8760, 2019.
Velickovic, P., Cucurull, G., et al. Graph Attention Net- Xiong, Z., Wang, D., and others. Pushing the boundaries of
works. In ICLR, 2018. molecular representation for drug discovery with graph
attention mechanism. J Med Chem, 2020.
Wang, S., Guo, Y., and others. Smiles-bert - large scale un-
supervised pre-training for molecular property prediction. Yang, K., Swanson, K., and others. Analyzing learned
BCB, 2019. molecular representations for property prediction. J
CHEM INF MODEL, 2019.
Wang, Y., Bryant, S. H., et al. Pubchem Bioassay: 2017
Update. Nucleic Acids Res., 2017. Yap, C. W. Padel-descriptor: An open source software to
calculate molecular descriptors and fingerprints. Journal
Weininger, D., Weininger, A., and Weininger, L. J. Smiles. of computational chemistry, 32(7):1466–1474, 2011.
2. algorithm for generation of unique smiles nota-
tion. JOURNAL OF CHEMICAL INFORMATION Ying, C., Cai, T., et al. Do Transformers Really Perform
AND COMPUTER SCIENCES, 1989. Badly for Graph Representation? In NeurIPS, 2021.
Wu, Z., Ramsundar, B., Feinberg, E. N., Gomes, J., Ge- You, Y., Chen, T., and others. Graph contrastive learning
niesse, C., Pappu, A. S., Leswing, K., and Pande, V. with augmentations. In NeurIPS, 2020.
Moleculenet: a benchmark for molecular machine learn- Yue, L., Jun, X., Sihang, Z., Siwei, W., Xifeng, G., Xi-
ing. Chemical science, 9(2):513–530, 2018. hong, Y., Ke, L., Wenxuan, T., Wang, L. X., et al. A
survey of deep graph clustering: Taxonomy, challenge,
Xia, J., Lin, H., et al. Towards Robust Graph Neural
and application. arXiv preprint arXiv:2211.12875, 2022.
Networks against Label Noise, 2021. URL https:
//openreview.net/forum?id=H38f_9b90BO. Yüksel, A., Ulusoy, E., Ünlü, A., Deniz, G., and Doğan,
T. Selformer: Molecular representation learning via self-
Xia, J., Tan, C., Wu, L., Xu, Y., and Li, S. Z. Ot cleaner: ies language models. arXiv preprint arXiv:2304.04662,
Label correction as optimal transport. In ICASSP 2023.
2022-2022 IEEE International Conference on Acoustics,
Speech and Signal Processing (ICASSP), pp. 3953–3957. Zernov, V. V., Balakin, K. V., Ivaschenko, A. A., Savchuk,
IEEE, 2022a. N. P., and Pletnev, I. V. Drug discovery using sup-
port vector machines. the case studies of drug-likeness,
Xia, J., Wu, L., , Chen, J., Hu, B., and Li, S. Z. SimGRACE: agrochemical-likeness, and enzyme inhibition predictions.
A Simple Framework for Graph Contrastive Learning Journal of chemical information and computer sciences,
without Data Augmentation. In Proceedings of The Web 43(6):2048–2056, 2003.
Conference 2022. Association for Computing Machinery,
2022b. Zheng, J., Wang, Y., Wang, G., Xia, J., Huang, Y., Zhao,
G., Zhang, Y., and Li, S. Z. Using context-to-vector with
Xia, J., Wu, L., et al. ProGCL: Rethinking Hard Negative graph retrofitting to improve word embeddings. arXiv
Mining in Graph Contrastive Learning. In ICML, 2022c. preprint arXiv:2210.16848, 2022.
Xia, J., Zheng, J., Tan, C., Wang, G., and Li, S. Z. Towards Zheng, S., Yan, X., Yang, Y., and Xu, J. Identify-
effective and generalizable fine-tuning for pre-trained ing structure–property relationships through smiles syn-
molecular graph models. bioRxiv, 2022d. tax analysis with self-attention mechanism. Journal
7
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
8
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
AUC_ROC
AUC_ROC
1.2
RMSE
0.88 0.82
1.0 0.86 0.81
0.84 0.80
0.8 0.82 0.79
0.80 0.78
Original Transformed Original Transformed Original Transformed
Figure 4. The performance of various models on the orthogonally transformed datasets. Left: FreeSolv (Regression); Middle: ClinTox
(Classification); Right: Tox21 (Classification). Kindly note that we did not evaluate CNN, RNN, and TRSF on the transformed datasets
because we cannot apply the orthogonal transformations to the input SMILES strings.
B. Experimental Setups
Fingerprints 7−→ SVM, XGB, RF, and MLP. Following the common practice (Tian et al., 2022; Pattanaik & Coley,
2020), we feed the concatenation of various molecular fingerprints including 881 PubChem fingerprints (PubchemFP),
307 substructure fingerprints (SubFP), and 206 MOE 1-D and 2-D descriptors (Yap, 2011) to SVM, XGB, RF, and MLP
models to comprehensively represent molecular structures, with some pre-processing procedures to remove features (1)
with missing values; (2) with extremely low variance (variance < 0.05); (3) have a high correlation (Pearson correlation
coefficient > 0.95) with another feature. The retained features are normalized to the mean value of 0 and variance of 1.
Additionally, considering that traditional machine models (SVM, RF, XGB) cannot be directly applied in the multi-task
molecular datasets, we split the multi-task dataset into multiple single-task datasets and use each of them to train the models.
Finally, we report the average performance of these single tasks.
SMILES strings 7−→ CNN, RNN, and TRSF. We adopt the 1D CNNs from a recent study (Kimber et al., 2021), which
include a single 1D convolutional layer with a step size equal to 1, followed by a fully connected layer. As for the RNN, we
use a 3-layer bidirectional gated recurrent units (GRUs) (Cho et al., 2014) with 256 hidden vector dimensions. Additionally,
we use the pre-trained SMILES transformer (Honda et al., 2019) with 4 basic blocks and each block has 4-head attentions
with 256 embedding dimensions and 2 linear layers. The SMILES are split into symbols (e.g., ‘Br’, ‘C’, ‘=’, ‘(’,‘2’) and
then fed into the transformer together with the positional encoding (Vaswani et al., 2017).
2D Graphs 7−→ GCN, MPNN, GAT, and AFP. As in previous studies (Xiong et al., 2019), we exhaustively utilized all
readily available atom/bond features in our 2D graph-based descriptors. Specifically, we have incorporated 9 atom features,
including atom symbol, degree, and formal charge, using a one-hot encoding scheme. In addition, we included 4 bond
features, such as type, conjugation, ring, and stereo. The resulting encoded graphs were then fed into GCN, MPNN, GAT,
and AFP models. Further details on the graph descriptors used in our experiments can be found in (Xiong et al., 2019).
3D Graphs 7−→ SPN. We employ the recently proposed SphereNet (Liu et al., 2022) for molecules with 3D geometry.
Specifically, for quantum mechanics datasets (QM7 and QM8) that contain 3D atomic coordinates calculated with ab initio
Density Functional Theory (DFT), we feed them into SphereNet directly. For other datasets without labeled conformations,
we used RDKit (Landrum, 2013)-generated conformations to satisfy the request of SphereNet.
Datasets splits, evaluation protocols and metrics, hyper-parameters tuning. Firstly, we randomly split the training,
validation, and test sets at a ratio of 8:1:1. And then, we tune the hyper-parameters based on the performance of the validation
set. Specifically, we select the optimal hyper-parameters set using the Tree of Parzen Estimators (TPE) algorithm (Ozaki
et al., 2020) in 50 evaluations. Due to the heavy computational overhead, GNNs-based models on the HIV and MUV datasets
are in 30 evaluations; all the models on the QM7 and QM8 are in 10 evaluations. And then, we conduct 50 independent
9
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
runs with different random seeds for dataset splitting to obtain more reliable results, using the optimal hyper-parameters
determined before. Similarly, GNNs-based models on the HIV and MUV datasets are in 30 evaluations; all the models on
the QM7 and QM8 are in 10 evaluations. Following MoleculeNet benchmark (Wu et al., 2018), we evaluate the classification
tasks using the area under the receiver operating characteristic curve (AUC-ROC), except the area under the precision curve
(AUC-PRC) on MUV dataset due to its extreme biased data distribution. The performance on the regression task are reported
using root mean square error (RMSE) or mean absolute error (MAE). kindly note that we report the average performance
across multi-tasks on some datasets because they contain more than one task. Additionally, to avoid the overfitting issue,
all the deep models are trained with an early stopping scheme if no validation performance improvement is observed in
successive 50 epochs. We set the maximal epoch as 300 and the batch-size as 128.
C. Related Work
In this section, we elaborate on various molecular descriptors and their respective learning models.
10
Why Deep Models Often Cannot Beat Non-deep Counterparts on Molecular Property Prediction?
In addition, it is also possible to directly leverage stereochemistry information such as chirality given the 3D geometries.
Recently, multiple works (Schuett et al., 2017; Satorras et al., 2021; Du et al., 2022; Liu et al., 2022; Atz et al., 2021) have
developed message-passing mechanisms tailored for 3D geometries, which enable the learned molecular representations
to follow certain physical symmetries, such as equivariance to translations and rotations. However, the calculation cost,
alignment invariance, uncertainty in conformation generation, and unavailable conformations of target molecules limited the
applicability of these models in practice.
11