Next Article in Journal
Synthesis and Structural and Optical Behavior of Dehydrohelicene-Containing Polycyclic Compounds
Next Article in Special Issue
Recent Progress of Protein Tertiary Structure Prediction
Previous Article in Journal
Aromas: Lovely to Smell and Nice Solvents for Polyphenols? Curcumin Solubilisation Power of Fragrances and Flavours
Previous Article in Special Issue
Naive Prediction of Protein Backbone Phi and Psi Dihedral Angles Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quantitative Predictive Studies of Multiple Biological Activities of TRPV1 Modulators

1
State Key Laboratory of NBC Protection for Civilian, Beijing 102205, China
2
School of Physics and Electronic Engineering, Sichuan University of Science & Engineering, Zigong 643000, China
*
Authors to whom correspondence should be addressed.
Molecules 2024, 29(2), 295; https://doi.org/10.3390/molecules29020295
Submission received: 3 December 2023 / Revised: 25 December 2023 / Accepted: 27 December 2023 / Published: 5 January 2024

Abstract

:
TRPV1 channel agonists and antagonists, which have powerful analgesic effects without the addictive qualities associated with traditional analgesics, have become a focus area for the development of novel analgesics. In this study, quantitative structure–activity relationship (QSAR) models for three bioactive endpoints (Ki, IC50, and EC50) were successfully constructed using four machine learning algorithms: SVM, Bagging, GBDT, and XGBoost. These models were based on 2922 TRPV1 modulators and incorporated four types of molecular descriptors: Daylight, E-state, ECFP4, and MACCS. After the rigorous five-fold cross-validation and external test set validation, the optimal models for the three endpoints were obtained. For the Ki endpoint, the Bagging-ECFP4 model had a Q2 value of 0.778 and an R2 value of 0.780. For the IC50 endpoint, the XGBoost-ECFP4 model had a Q2 value of 0.806 and an R2 value of 0.784. For the EC50 endpoint, the SVM-Daylight model had a Q2 value of 0.784 and an R2 value of 0.809. These results demonstrate that the constructed models exhibit good predictive performance. In addition, based on the model feature importance analysis, the influence between substructure and biological activity was also explored, which can provide important theoretical guidance for the efficient virtual screening and structural optimization of novel TRPV1 analgesics. And subsequent studies on novel TRPV1 modulators will be based on the feature substructures of the three endpoints.

Graphical Abstract

1. Introduction

TRPV1 channels are nociceptors found on C and Aδ fibers [1]. They detect various noxious stimuli, such as high temperatures (>42 °C), acidity (H+), and a range of endogenous and exogenous ligands [2]. These channels are crucial in pain management. TRPV1 modulators, including agonists and antagonists, have demonstrated significant efficacy in the treatment of neuropathic pain, osteoarthritis, and cancer pain [3]. Among them, TRPV1 agonists produce long-lasting and reversible analgesia through calcium-dependent desensitization, rendering TRPV1-expressing nerve fibers unresponsive to noxious stimuli [4]. TRPV1 antagonists, on the other hand, reduce nociceptive hypersensitivity by inhibiting TRPV1 channels, thus inhibiting the production of noxious sensations. Traditional analgesics, such as opioid narcotic analgesics and nonsteroidal anti-inflammatory analgesics, while initially providing temporary or partial pain relief, are associated with dose-limiting side-effects, lack of tolerance, and decreased efficacy over time, particularly impacting the treatment of chronic pain in the elderly [5]. However, existing TRPV1 modulators have side effects like strong irritation and hyperthermia, limiting their long-term clinical application [6,7]. Therefore, it is still necessary to develop novel TRPV1 modulators.
The agonistic or inhibitory activity of TRPV1 modulators is generally quantified using the concentration for 50% of maximal effect (EC50) or half maximal inhibitory concentration (IC50). Ki is the inhibition constant, which is a more precise indicator than IC50. The experimental approach to detect the effect of TRPV1 regulators on the opening (agonism) or closing (antagonism) of TRPV1 channels commonly involves the use of FLIPR [8] and electrophysiological membrane clamp [9]. Although intuitively clear, these experimental methods require the synthesis of the compound to be tested first and later assayed on TRPV1-expressing cells. Moreover, the speed of drug discovery is limited by the experimental methods, although high-throughput screening and combinatorial chemistry have been developed. Both the in vitro experiments, especially electrophysiological assays, require significant time and high investment costs.
The discovery of hits is a mandatory pathway to the discovery of novel TRPV1 modulators. However, high-throughput screening based on wet assays, combinatorial chemistry, and fragment-based drug design requires significant labor, material, and time costs and consumes a lot of effort on inactive compounds [10,11]. In recent years, computer-aided drug design, represented by quantitative structure–activity relationships (QSARs), has been rapidly developed due to the rise of artificial intelligence and big data [11]. QSAR modeling is a mathematical or statistical methodology that establishes a quantitative mapping between molecular structure and biological activity that can be used to predict the biological activity of new compounds on specific targets [11]. This methodology has been widely used in the discovery of various drug hits. There have been some structural modification studies of TRPV1 regulators based on 3D-QSAR models. For instance, Kristam et al. [12] constructed a 3D-QSAR model using 62 piperazine-aryl derived TRPV1 compounds with good predictive performance (Q2 = 0.9, R2 = 0.75). Then, they used a Topomer-CoMFA method to construct a new 3D-QSAR model [13]. However, the predictive performance of the new model did not improve. Similarly, Wang et al. [14] constructed a 3D-QSAR model with good predictive performance (Q2 = 0.522, R2 = 0.839) using the CoMSIA method based on 236 TRPV1 antagonists. Although these 3D-QSAR models showed promising results, they require the superposition of molecular 3D conformations. Unfortunately, the effect of conformation overlap of 3D-QSAR method could seriously affect the robustness of the models. Furthermore, 3D-QSAR models are often limited to predicting the properties of compounds with similar structures, thus having poor generalization ability [15].
To address the above problems and build a model with good generalization ability and stability, this study successfully constructed several QSAR models based on multiple machine learning algorithms for the three activity endpoints (EC50 of TRPV1 agonists, IC50 of TRPV1 antagonists, and Ki). The internal and external validation showed that the models have good predictive performance and generalization ability, which can provide high-quality virtual screening models for the development of novel TRPV1 modulators.

2. Results and Discussion

2.1. Chemical Space and Scaffold Analysis

To construct the QSAR model, it is important for the dataset to encompass a wide range of activity. The Ki dataset ranges from 5.76 to 10.00 in terms of pKi values, the EC50 dataset ranges from 3.95 to 8.72 in terms of pEC50 values, and the IC50 dataset ranges from 4.04 to 9.40 in terms of pIC50 values. Therefore, the datasets for the three activity endpoints cover a broad span of activity, ranging from µM to nM. The activity distributions of the training and testing sets for the three activity endpoints, as indicated by the histograms (Figure 1A–C), closely resemble those of the total dataset. This suggests that the division of the dataset is reasonable with respect to activity distribution. In addition, principal component analysis (PCA) was utilized to represent the scaffold distribution of the compounds in both the training and test sets (Figure 1). Notably, the compound scaffolds representing the test set of the three endpoints were mainly distributed within the compound scaffolds of their corresponding training sets, and no more outliers appeared. Hence, the scaffold division of the test set proved suitable for evaluating the predictive performance and generalization capability of the QSAR model.
Table 1 lists the top ten carbon scaffolds with the highest numbers, the vast majority of which include an isobutane carbon scaffold structure corresponding to the neck group of the TRPV1 modulator, typically an amide, ureido, or thiourea, among others. The analysis of compound structures in the datasets showed 77 carbon scaffolds in the Ki dataset, 275 in the IC50 dataset, and 97 in the EC50 dataset. This indicates a significant diversity in structural composition across the datasets for the three endpoints. In contrast, the head and tail in the backbone are mostly cyclic structures, corresponding to the tail moiety that forms a hydrophobic interaction in TRPV1 modulators and the head moiety that mostly contains an aromatic ring. It is noteworthy that some special carbon scaffolds appear in the scaffold of EC50. First, the cyclohexane carbon scaffold, ranked in terms of content, is quite different from the generic structure of TRPV1 modulators, and can be designed as a head group to provide ideas for fragment-based drug design. Secondly, the carbon scaffolds ranked fifth and sixth in terms of content both appear to have a bridging ring structure and can be designed as a tail moiety that can provide strong van der Waals interactions and help increase binding affinity [16].

2.2. Feature Selection

To enhance the interpretability and accuracy of the model while minimizing training costs, a method known as recursive feature elimination based on random forest (RFE-RF) is employed for feature selection. Initially, RFE-RF utilizes the complete set of features from a descriptor or molecular fingerprint for modeling. Subsequently, it proceeds to eliminate the least significant feature iteratively, employing the remaining features for subsequent modeling steps. Finally, it selects the combination of features with the lowest RMSECV.
As shown in Figure 2, the performance of the model gradually improves as the number of features increases, eventually reaching a plateau. The number of features selected varies across the three active endpoints for different descriptors or molecular fingerprints. The red dots identified by orange dashed lines in each subplot of Figure 2 indicate the selected feature combinations. In Ki, the optimal number of features for Daylight, E-state, ECFP4, and MACCS was 53 (2.6% of original features), 30 (27.3% of original features), 82 (8.0% of original features), and 50 (30.1% of original features), respectively; in IC50, the optimal number of features for Daylight, E-state, ECFP4, and MACCS was 183 (8.9% of original features), 33 (30.0% of original features), 293 (28.6% of original features), and 68 (41.0% of original features), respectively; and in EC50, Daylight, E-state, ECFP4, and MACCS have the optimal number of features of 168 (8.2% of original features), 25 (22.7% of original features), 55 (5.4% of original features), and 22 (13.3% of original features), respectively. These 12 sets of features will be used as independent variables to construct 48 active prediction models using four machine learning algorithms, i.e., 16 models for each endpoint.

2.3. Evaluation of Ki Activity Prediction Models

The evaluation results of the 16 Ki activity prediction models are presented in Table 2. It can be observed that the internal validation results of different algorithms under the same descriptor are similar. Furthermore, there is a consistent trend in the internal validation results of different descriptors under the same algorithm, with ECFP4 showing the highest performance, followed by Daylight, MACCS, and E-state. The model constructed using the Bagging algorithm and ECFP4 descriptors demonstrates the highest performance (Q2 = 0.778, R2 = 0.780), while the models constructed using SVM and E-state descriptors exhibit the lowest performance (Q2 = 0.502, R2 = 0.536). The MAECV and MAET of the vast majority of the models were in the range of 0.3–0.4, indicating that the difference between the predicted results and the experimental values was not more than half an order of magnitude. Thus, the predicted values of these models are of practical significance. Subsequently, the external validation results of the 16 models align with the internal validation results, reaffirming their good generalization ability and reliable prediction capability for the Ki activity values of new chemical entities. Figure 3 displays the scatter plot of the predicted values of the optimal model against the experimental values. The green dots represent the training set, while the orange dots represent the test set.

2.4. Evaluation of IC50 Activity Prediction Models

Table 3 presents the evaluation results of 16 IC50 activity prediction models. Among SVM, Bagging, and XGBoost, the prediction performance of the four descriptors is ranked as follows: ECFP4 > Daylight > MACCS > E-state. However, in GBDT, the prediction performance of Daylight is stronger than ECFP4. Comparing the internal validation results of Ki with those of the four algorithms in the IC50 dataset, there are significant differences. Specifically, the internal validation performance of GBDT is significantly lower than that of the other three algorithms. It is worth noting that, although the SVM model using E-state descriptors has the worst internal validation performance (Q2 = 0.487 ± 0.008, RMSECV = 0.424 ± 0.005, and MAECV = 0.338 ± 0.004) among all the models, the model constructed by XGBoost and ECFP4 is considered the optimal model for IC50 activity prediction. This model performs the best for both internal and external validation, demonstrating good generalization performance. Furthermore, the MAE of the IC50 model is comparable to that of the Ki prediction model, with errors within half an order of magnitude. Figure 4 illustrates the scatterplot of predicted versus experimental values for the optimal model, with green dots indicating the training set and orange dots representing the test set. The green solid line represents the trend line for the training set, while the orange solid line corresponds to the trend line for the test set. Notably, the trend lines of the training and test sets resemble those of the Ki prediction model, further confirming the model’s strong generalization ability.

2.5. Evaluation of EC50 Activity Prediction Models

Table 4 presents the performance evaluation results for the 16 EC50 activity prediction models. It is evident that the internal validation of the same descriptors varies less across different algorithms. However, the performance of the E-state descriptor in the SVM model is noticeably inferior to the other three algorithms. Additionally, the performance of the four descriptors in the Bagging, GBDT, and XGBoost algorithms follows the order of ECFP4 > Daylight > MACCS > E-state. Conversely, in the SVM algorithm, the internal validation of Daylight outperforms that of ECFP4, making it the optimal model among the EC50 activity prediction models with an external validation R2 exceeding 0.8, thus highlighting its exceptional predictive capability. Figure 5 illustrates the scatter plots comparing the predicted and experimental values for the SVM and Daylight models.
The results of internal and external validation for our models demonstrate a significantly reduced difference (less than 0.03) between Q2 and R2 compared to the previous 3D-QSAR model [12,13,14] (Table 5), which exhibited a difference of more than 0.25. This indicates a significant improvement in the generalization ability of the model. The enhanced performance can be attributed, primarily, to the larger dataset utilized in this study, as well as the robust stability of the machine learning algorithms employed. Notably, the model proposed by Kristam was developed using only 62 molecules, making it challenging to ensure generalizability.

2.6. Y-Randomization Test

Feature selection involves selecting the best performing feature combinations from high-dimensional descriptors and molecular fingerprints to build models that are highly fitted to experimental values. However, it is possible to obtain such models by chance, without any real correlation between the descriptors and experimental values. To assess the chance correlation of the model, we applied the Y-randomization test. During the Y-randomization test, the experimental values of pKi, pIC50, and pEC50 are randomly disrupted, destroying the original relationship between the descriptors or molecular fingerprints and the activity values, but the distribution of the activity values does not change [17]. We then re-modeled the disrupted data using the algorithm of the three optimal models and the molecular fingerprints, repeating the process 1000 times. The results of evaluating the 1000 randomized models using Q2 are shown in Figure 6. In this figure, the horizontal coordinate represents Q2, and the vertical coordinate represents the number of frequencies. The green bars on the left side of the three subfigures represent the histograms of the distribution of Q2 for the 1000 randomized models, while the orange vertical lines indicate the Q2 of the original models. From Figure 6, it is evident that all the Q2 of the randomized models fall between −1 and 0, indicating no correlation between the true value of the randomized model and the descriptor or molecular fingerprints. According to the paired-sample t-test, the confidence level of the randomized model compared to the original model is 99% (p < 0.001), which is statistically significant. Therefore, in the three optimal activity prediction models constructed in this paper, there exists a real correlation, rather than a chance correlation, between the modeled molecular fingerprints and Ki, IC50, or EC50.

2.7. Model Interpretation

In this paper, we aim to interpret the three optimal models by ranking the importance of features. To select the features, we employ the RFE-RF method, which calculates the Gini index for each feature to indicate its significance. Figure 7 displays the five features with the highest importance among the three optimal models. Additionally, it is worth mentioning that the optimal models of Ki and IC50 utilize the ECFP4 fingerprint, whereas the optimal model of EC50 utilizes the Daylight fingerprint. Notably, the sum of the importance of all the features in the models is equal to 1.
The first five features of Ki had a cumulative importance of 0.493. Out of the dataset, 349 compounds had these features in the following descending order: 200, 667, 573, 316, and 997. This accounted for 52.80% of the total Ki dataset. Figure 8A displays the histogram of pKi distribution for compounds containing the top 5 features. It is evident that the pKi of these compounds is shifted one unit to the right compared to other compounds. The structure of the first five features is shown in Figure 9A. The structure of TRPV1 modulators typically comprises three parts: the head, the neck, and the tail. Generally, the head serves as a hydrogen bond acceptor, while the neck acts as a hydrogen bond acceptor and is commonly an amide, urea, or thiourea. On the other hand, the tail is a hydrophobic group [18,19]. In the case of these compounds, the first five characteristics correspond to the head (positions 667, 316, 997) and neck (positions 200, 573), with the head being a methylsulfonamide attached to a benzene ring and the neck being an amide group. Referring to the activity distribution in Figure 8A, it is reasonable to assume that compounds containing such a structure tend to exhibit high Ki activity and hold potential for modification.
The first three features of IC50 have been found to be significantly more important than the last two, with the order of importance being 672 bits > 128 bits > 378 bits. Therefore, the first three features are selected for model interpretation, as illustrated in Figure 7. In the IC50 dataset, there were 273 compounds (14.41% of the dataset) that demonstrated the first three features. The cumulative importance of these features amounted to 0.225. Figure 8B highlights that compounds possessing the first three features exhibited a rightward shift in pIC50 compared to other compounds, suggesting a higher level of antagonistic activity. As illustrated in Figure 9B, positional markers 672 and 128 correspond to the aromatic ring of the head and the urea group of the neck, respectively, while position 378 signifies the indole ring of the head. The indole in the head acts as both a hydrogen bond donor and acceptor, facilitating specific interactions for antagonist binding to TRPV1 channels. Therefore, compounds that incorporate 1H-indole in the head may potentially possess highly active antagonistic properties.
The Daylight fingerprint is different from ECFP4 in that it represents the molecular structure as a linear path from atoms, and thus has no central atom. The importance of the 67-position feature is 0.311, which is much higher than that of the other features, and the number of compounds with this feature is 205, which accounts for 55.86% of the EC50 dataset, thus this feature is used to interpret the model. In Figure 8C, the compounds with position 67 have a significant rightward shift in pEC50 compared to the other compounds, and the IC50 activity is nearly two orders of magnitude different. As can be seen in Figure 9C, the 67-position feature indicates the ureido group in the neck and the benzene ring in the head. This indicates that most of the highly active TRPV1 agonists have phenylurea at the neck and head, and thus compounds containing phenylurea are potentially highly active TRPV1 agonists.

3. Materials and Methods

3.1. Data Collection and Processing

Data on the Ki, IC50, and EC50 activities of human TRPV1 channels were collected from the ChEMBL [20] and PubChem [21] databases. The data were processed according to the following steps: 1. compounds without a clear type of activity and activity value were removed; 2. the units of nanomoles (nM) or micromoles (μM) in the original data were converted to M and the negative logarithms with a base of 10 were taken (i.e., pKi, pIC50, and pEC50); 3. the compounds with multiple activity values were de-weighted, following the rule that if the maximum difference of the negative logarithm is less than or equal to 1, the mean value is taken as the activity value of the compound, and the compound is discarded otherwise; 4. salt ions and metal ions were removed from the dataset. After processing, three activity datasets were obtained, consisting of 661 Ki, 1894 IC50, and 367 EC50 values. Based on these datasets, three QSAR models were constructed.

3.2. Descriptor Generation

This paper utilizes four different methods to extract features and build QSAR models: Daylight fingerprints, molecular access system (MACCS) [22] fingerprints, electrotopological state indices (E-state) molecular descriptors [23], and Extended-Connectivity Fingerprints (ECFPs) [24]. MACCS fingerprints are substructure-based molecular fingerprints, and this paper selects the commonly used 166-bit fingerprints. Daylight fingerprints, also known as path-based molecular fingerprints, characterize molecules through different atomic paths represented by a total of 2048 bits. E-state molecular descriptors simultaneously characterize the molecular structure and electrical characteristics with a total of 110 features. ECFP fingerprint is a circular topological fingerprint based on Morgan’s algorithm. This study uses ECFP4 with a diameter of 4 and 1024 bits. These descriptors of compounds in databases were calculated through the Scopy [25] and rdkit [26] toolkit.

3.3. Data Set Segmentation

In order to avoid training bias or overfitting and to maintain similar structural distribution of compounds in each subset close to each other, this paper divides the dataset into training and test sets according to the carbon scaffold. The carbon scaffold is determined by removing all R groups from the molecule and retaining only the connecting groups between the ring systems, while converting heteroatoms to carbon atoms and bonding sequences to single bonds. The Scopy toolkit [25] is employed to calculate the carbon scaffold of the compounds in this study. If there are less than five molecules with the same carbon scaffold, one molecule is randomly assigned to the training set and the remaining molecules are assigned to the test set. On the other hand, if there are five or more molecules with the same carbon scaffold, 80% of them are randomly assigned to the training set while the remaining 20% are assigned to the test set.

3.4. Machine Learning Methods

In this study, four machine learning algorithms are used for the construction of QSAR models, namely support vector machine (SVM), gradient boosting decision tree (GBDT), extreme gradient boosting (XGBoost) and bagging. The models of the 4 algorithms were implemented via the scikit-learn toolkit [27].
SVM [28] is a statistical learning algorithm based on the principle of Vapnik structural risk minimization. Originally developed for classification problems, SVM can also be extended to regression tasks by introducing slack variables. In the regression task, the objective is to find a hyperplane with a small number of paradigms, while minimizing the sum of the distances from the data to the hyperplane [29]. Its high degree of generalization ability has contributed to its increasing popularity in the QSAR/QSPR species.
GBDT [30] is a machine learning algorithm based on the idea of Boosting integration. GBDT updates the strong learner by decreasing the loss function, fitting the loss approximation for each round of iteration with the negative gradient of the loss function. A disadvantage of GBDT is that it is difficult to train in parallel and is less efficient.
XGBoost was developed by Tianqi Chen et al. [31]. Other Boosting algorithms develop their models in a sequential phase manner like other Boosting algorithms. However, XGBoost enables parallel computation and also has improved handling of missing values compared to GBDT. In addition, XGBoost is highly resistant to overfitting due to the inclusion of regular terms.
Unlike Boosting, each base learner in Bagging [32] is independent and can be computed in parallel. Bagging samples n sample sets using an autonomous sampling method and training a base learner for each sample set. Afterwards, the learners are combined. Hence, approximately 36.8% of the samples in the initial dataset do not appear in the sampling set. These samples can be used as a validation set to test the training performance and generalization ability of the model. The Bagging algorithm focuses on the reduction in variance and is known for its integration and efficiency.
The grid search in scikit-learn was used for parameter tuning. The key parameters of SVM are C (the penalty coefficient) and gamma (the coefficient of the kernel function). In grid search, the values of C were set as 0.01, 0.1, 1, 10, 100, and 1000; the values of gamma were set as 0.0001, 0.001, and 0.01; and the kernel was chosen as RBF. The number of decision trees is the parameter of Bagging, XGBoost, and GBDT ranging from 100 to 1000, with a step size of 50.

3.5. Performance Evaluation Indicators

To ensure the good generalization ability of the QSAR model in predicting the biological activity of new chemical entities, internal validation and external validation were conducted. The model was internally validated using five-fold cross-validation (CV) and independent test sets. In five-fold CV, the training set was divided into five equal parts, with four parts used for constructing the model and one part used for model validation. This process was repeated five times, allowing each part of the data to serve as a validation set. Four main statistical parameters were employed to evaluate the model’s performance: the coefficient of determination (Q2), the root mean square error (RMSECV), and the mean absolute error (MAECV) for CV and the coefficient of determination (R2), the root mean square error (RMSET), and the mean absolute error (MAET) for the test set. The formulas for Q2 (R2), RMSE, and MAE are given below:
Q 2 R 2 = 1 i = 1 n ( y ^ i y i ) 2 i = 1 n ( y ¯ i y i ) 2
R M S E = 1 n i = 1 n y ^ i y i 2
M A E = 1 n i = 1 n y ^ i y i
where y ^ i is the predicted value, y i is the true value, and y ¯ i is the average value of y i in the sample. From the formula, it can be seen that the smaller the value of RMSE and MAE, the better the performance of the model; while the larger the value of Q2 or R2, the better the performance of the model.

4. Conclusions

To accelerate the discovery of novel TRPV1 modulators, QSAR models that can quantitatively predict Ki, IC50, and EC50 were constructed using four machine learning algorithms based on 2922 biological activity data. After rigorous internal and external validation, the constructed models exhibited excellent external prediction performance and generalization ability. The model feature importance analysis revealed that the key feature structures of the three endpoints were concentrated in the head and neck of the molecule, aligning with the conclusion that the polar interactions between the TRPV1 regulator and TRPV1 only existed in their head and neck region. Specifically, a higher Ki activity tended to be observed in molecules with a methylsulfonamide attached to a benzene ring in the head and an amide group in the neck. Additionally, molecules containing 1H-indole in the head showed potential as highly active antagonists, while those containing phenylurea have a likely potential to be highly active TRPV1 agonists. These findings pertaining to the influence of the microstructure of TRPV1 modulators on their biological activities are expected to provide guidance for the rational design and efficient screening of novel analgesic drugs.

Author Contributions

Conceptualization, X.W. and L.W.; methodology, T.H. and Z.Y.; software, T.H.; validation, X.W., L.P. and L.W.; formal analysis, X.W.; investigation, X.W.; resources, J.D.; data curation, T.H. and Z.Y.; writing—original draft preparation, X.W.; writing—review and editing, L.W. and J.D.; visualization, T.H.; supervision, L.W. and L.P.; project administration, J.D.; funding acquisition, J.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Caterina, M.J.; Julius, D. The vanilloid receptor: A molecular gateway to the pain pathway. Annu. Rev. Neurosci. 2001, 24, 487–517. [Google Scholar] [CrossRef]
  2. Bevan, S.; Quallo, T.; Andersson, D.A. TRPV1. In Mammalian Transient Receptor Potential (TRP) Cation Channels: Volume I; Nilius, B., Flockerzi, V., Eds.; Springer: Berlin/Heidelberg, Germany, 2014; pp. 207–245. [Google Scholar]
  3. Iftinca, M.; Defaye, M.; Altier, C. TRPV1-Targeted Drugs in Development for Human Pain Conditions. Drugs 2021, 81, 7–27. [Google Scholar] [PubMed]
  4. Moran, M.M. TRP Channels as Potential Drug Targets. Annu. Rev. Pharmacol. Toxicol. 2018, 58, 309–330. [Google Scholar] [CrossRef] [PubMed]
  5. Domenichiello, A.F.; Ramsden, C.E. The silent epidemic of chronic pain in older adults. Prog. Neuropsychopharmacol. Biol. Psychiatry 2019, 93, 284–290. [Google Scholar] [PubMed]
  6. Gladkikh, I.N.; Sintsova, O.V.; Leychenko, E.V.; Kozlov, S.A. TRPV1 Ion Channel: Structural Features, Activity Modulators, and Therapeutic Potential. Biochemistry 2021, 86, S50–S70. [Google Scholar] [PubMed]
  7. Garami, A.; Shimansky, Y.P.; Rumbus, Z.; Vizin, R.C.L.; Farkas, N.; Hegyi, J.; Szakacs, Z.; Solymar, M.; Csenkey, A.; Chiche, D.A.; et al. Hyperthermia induced by transient receptor potential vanilloid-1 (TRPV1) antagonists in human clinical trials: Insights from mathematical modeling and meta-analysis. Pharmacol. Ther. 2020, 208, 107474. [Google Scholar] [PubMed]
  8. Moriello, A.S.; De Petrocellis, L.; Vitale, R.M. Fluorescence-Based Assay for TRPV1 Channels. In Endocannabinoid Signaling: Methods and Protocols; Maccarrone, M., Ed.; Springer: New York, NY, USA, 2023; pp. 119–131. [Google Scholar]
  9. Musella, A.; Centonze, D. Electrophysiology of Endocannabinoid Signaling. Methods Mol. Biol. 2023, 2576, 461–475. [Google Scholar]
  10. Li, Q.; Shah, S. Structure-Based Virtual Screening. Methods Mol. Biol. 2017, 1558, 111–124. [Google Scholar]
  11. Shaker, B.; Ahmad, S.; Lee, J.; Jung, C.; Na, D. In silico methods and tools for drug discovery. Comput. Biol. Med. 2021, 137, 104851. [Google Scholar]
  12. Kristam, R.; Parmar, V.; Viswanadhan, V.N. 3D-QSAR analysis of TRPV1 inhibitors reveals a pharmacophore applicable to diverse scaffolds and clinical candidates. J. Mol. Graph. Model. 2013, 45, 157–172. [Google Scholar]
  13. Kristam, R.; Rao, S.N.; D’Cruz, A.S.; Mahadevan, V.; Viswanadhan, V.N. TRPV1 antagonism by piperazinyl-aryl compounds: A Topomer-CoMFA study and its use in virtual screening for identification of novel antagonists. J. Mol. Graph. Model. 2017, 72, 112–128. [Google Scholar] [CrossRef] [PubMed]
  14. Wang, J.; Li, Y.; Yang, Y.; Du, J.; Zhang, S.; Yang, L. In silico research to assist the investigation of carboxamide derivatives as potent TRPV1 antagonists. Mol. Biosyst. 2015, 11, 2885–2899. [Google Scholar] [CrossRef] [PubMed]
  15. Melo-Filho, C.C.; Braga, R.C.; Andrade, C.H. 3D-QSAR approaches in drug design: Perspectives to generate reliable CoMFA models. Curr. Comput.-Aided Drug Des. 2014, 10, 148–159. [Google Scholar] [CrossRef] [PubMed]
  16. Yang, F.; Xiao, X.; Cheng, W.; Yang, W.; Yu, P.; Song, Z.; Yarov-Yarovoy, V.; Zheng, J. Structural mechanism underlying capsaicin binding and activation of the TRPV1 ion channel. Nat. Chem. Biol. 2015, 11, 518–524. [Google Scholar] [CrossRef]
  17. Rücker, C.; Rücker, G.; Meringer, M. y-Randomization and Its Variants in QSPR/QSAR. J. Chem. Inf. Model. 2007, 47, 2345–2357. [Google Scholar] [CrossRef]
  18. Szallasi, A.; Cortright, D.N.; Blum, C.A.; Eid, S.R. The vanilloid receptor TRPV1: 10 years from channel cloning to antagonist proof-of-concept. Nat. Rev. Drug Discov. 2007, 6, 357–372. [Google Scholar] [CrossRef]
  19. Aghazadeh Tabrizi, M.; Baraldi, P.G.; Baraldi, S.; Gessi, S.; Merighi, S.; Borea, P.A. Medicinal Chemistry, Pharmacology, and Clinical Implications of TRPV1 Receptor Antagonists. Med. Res. Rev. 2017, 37, 936–983. [Google Scholar] [CrossRef]
  20. Mendez, D.; Gaulton, A.; Bento, A.P.; Chambers, J.; De Veij, M.; Félix, E.; Magariños, M.P.; Mosquera, J.F.; Mutowo, P.; Nowotka, M.; et al. ChEMBL: Towards direct deposition of bioassay data. Nucleic Acids Res. 2019, 47, D930–D940. [Google Scholar] [CrossRef]
  21. Kim, S.; Chen, J.; Cheng, T.; Gindulyte, A.; He, J.; He, S.; Li, Q.; Shoemaker, B.A.; Thiessen, P.A.; Yu, B.; et al. PubChem in 2021: New data content and improved web interfaces. Nucleic Acids Res. 2021, 49, D1388–D1395. [Google Scholar] [CrossRef]
  22. Durant, J.L.; Leland, B.A.; Henry, D.R.; Nourse, J.G. Reoptimization of MDL Keys for Use in Drug Discovery. J. Chem. Inf. Comput. Sci. 2002, 42, 1273–1280. [Google Scholar] [CrossRef]
  23. Pearlman, R.S. Molecular Structure Description. The Electrotopological State by Lemont B. Kier (Virginia Commonwealth University) and Lowell H. Hall (Eastern Nazarene College). Academic Press:  San Diego. 1999. xx + 245 pp. $99.95. ISBN 0-12-406555-4. J. Am. Chem. Soc. 2000, 122, 6340. [Google Scholar] [CrossRef]
  24. Rogers, D.; Hahn, M. Extended-Connectivity Fingerprints. J. Chem. Inf. Model. 2010, 50, 742–754. [Google Scholar] [CrossRef] [PubMed]
  25. Yang, Z.-Y.; Yang, Z.-J.; Lu, A.-P.; Hou, T.-J.; Cao, D.-S. Scopy: An integrated negative design python library for desirable HTS/VS database design. Brief. Bioinform. 2020, 22, bbaa194. [Google Scholar] [CrossRef] [PubMed]
  26. Landrum, G.; Sforna, G.; Winter, H.D. RDKit: Open-Source Cheminformatics, version 2020.09; 2006. Available online: https://www.rdkit.org/ (accessed on 15 August 2023).
  27. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  28. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  29. Xu, J.; Wang, L.; Wang, L.; Shen, X.; Xu, W. QSPR study of Setschenow constants of organic compounds using MLR, ANN, and SVM analyses. J. Comput. Chem. 2011, 32, 3241–3252. [Google Scholar] [CrossRef]
  30. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  31. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: San Francisco, NY, USA, 2016; pp. 785–794. [Google Scholar]
  32. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
Figure 1. Distribution histograms (AC) and principal component analysis plots (DF) for the EC50, IC50, and Ki data sets.
Figure 1. Distribution histograms (AC) and principal component analysis plots (DF) for the EC50, IC50, and Ki data sets.
Molecules 29 00295 g001
Figure 2. Feature selection results based on RFE-RF. The red dot indicates the point with the lowest RMSE, and the horizontal and vertical dotted lines refer to the RMSE and the number of features corresponding to this point respectively.
Figure 2. Feature selection results based on RFE-RF. The red dot indicates the point with the lowest RMSE, and the horizontal and vertical dotted lines refer to the RMSE and the number of features corresponding to this point respectively.
Molecules 29 00295 g002
Figure 3. Scatter plot of predicted values versus experimental values of Ki prediction model based on Bagging and ECFP4. The green line indicates the trend line of the training set and the orange line indicates the trend line of the test set.
Figure 3. Scatter plot of predicted values versus experimental values of Ki prediction model based on Bagging and ECFP4. The green line indicates the trend line of the training set and the orange line indicates the trend line of the test set.
Molecules 29 00295 g003
Figure 4. Scatter plot of predicted values versus experimental values of IC50 prediction model based on XGBoost and ECFP4. The green line indicates the trend line of the training set and the orange line indicates the trend line of the test set.
Figure 4. Scatter plot of predicted values versus experimental values of IC50 prediction model based on XGBoost and ECFP4. The green line indicates the trend line of the training set and the orange line indicates the trend line of the test set.
Molecules 29 00295 g004
Figure 5. Scatter plot of predicted values versus experimental values of EC50 prediction model based on SVM and Daylight. The green line indicates the trend line of the training set and the orange line indicates the trend line of the test set.
Figure 5. Scatter plot of predicted values versus experimental values of EC50 prediction model based on SVM and Daylight. The green line indicates the trend line of the training set and the orange line indicates the trend line of the test set.
Molecules 29 00295 g005
Figure 6. The distribution of Q2 of randomization models of three activity endpoints. the left-side green bar represents the randomized Q2 distribution, and the orange vertical line on the right side represents the Q2 of the original model.
Figure 6. The distribution of Q2 of randomization models of three activity endpoints. the left-side green bar represents the randomized Q2 distribution, and the orange vertical line on the right side represents the Q2 of the original model.
Molecules 29 00295 g006
Figure 7. Descriptor importance of the top 5 features of 3 optimal models. The vertical coordinate represents the bit encoding of the molecular fingerprint, while the horizontal coordinate corresponds to the feature importance.
Figure 7. Descriptor importance of the top 5 features of 3 optimal models. The vertical coordinate represents the bit encoding of the molecular fingerprint, while the horizontal coordinate corresponds to the feature importance.
Molecules 29 00295 g007
Figure 8. (A) Histogram of activity distribution of compounds with and without feature structures for pKi endpoint; (B) Histogram of activity distribution of compounds with and without feature structures for pIC50 endpoint; (C) Histogram of activity distribution of compounds with and without feature structures for pEC50 endpoint.
Figure 8. (A) Histogram of activity distribution of compounds with and without feature structures for pKi endpoint; (B) Histogram of activity distribution of compounds with and without feature structures for pIC50 endpoint; (C) Histogram of activity distribution of compounds with and without feature structures for pEC50 endpoint.
Molecules 29 00295 g008
Figure 9. Feature structures and their position in the molecules. The blue structure on the left side of the molecule is the sum of the sub-structures corresponding to the features. On the right side, the featured structures are depicted, with the central atoms marked by purple dots. The atoms within the bonding radius are represented in black, while the green color is used to indicate the environments of the featured structures within the molecule. An asterisk indicates an unknown atom, which could be carbon, nitrogen, or something else. (A) The representative compound in Ki is shown on the left, and 5 substructures with the most importance are shown on the right. (B) The representative compound in IC50 and 3 substructures with the most importance. (C) The representative compound in EC50 and one substructure with the most importance.
Figure 9. Feature structures and their position in the molecules. The blue structure on the left side of the molecule is the sum of the sub-structures corresponding to the features. On the right side, the featured structures are depicted, with the central atoms marked by purple dots. The atoms within the bonding radius are represented in black, while the green color is used to indicate the environments of the featured structures within the molecule. An asterisk indicates an unknown atom, which could be carbon, nitrogen, or something else. (A) The representative compound in Ki is shown on the left, and 5 substructures with the most importance are shown on the right. (B) The representative compound in IC50 and 3 substructures with the most importance. (C) The representative compound in EC50 and one substructure with the most importance.
Molecules 29 00295 g009
Table 1. Top 10 carbon scaffolds and corresponding numbers of Ki, IC50, and EC50 data sets.
Table 1. Top 10 carbon scaffolds and corresponding numbers of Ki, IC50, and EC50 data sets.
No.KiIC50EC50
Carbon ScaffoldNumberCarbon ScaffoldNumberCarbon ScaffoldNumber
1Molecules 29 00295 i001137Molecules 29 00295 i002174Molecules 29 00295 i00341
2Molecules 29 00295 i00483Molecules 29 00295 i005152Molecules 29 00295 i00630
3Molecules 29 00295 i00747Molecules 29 00295 i008100Molecules 29 00295 i00924
4Molecules 29 00295 i01045Molecules 29 00295 i01177Molecules 29 00295 i01217
5Molecules 29 00295 i01337Molecules 29 00295 i01471Molecules 29 00295 i01514
6Molecules 29 00295 i01629Molecules 29 00295 i01763Molecules 29 00295 i01811
7Molecules 29 00295 i01922Molecules 29 00295 i02058Molecules 29 00295 i02111
8Molecules 29 00295 i02218Molecules 29 00295 i02346Molecules 29 00295 i02410
9Molecules 29 00295 i02518Molecules 29 00295 i02643Molecules 29 00295 i0279
10Molecules 29 00295 i02818Molecules 29 00295 i02935Molecules 29 00295 i0308
Table 2. The results of internal and external validation of Ki prediction models.
Table 2. The results of internal and external validation of Ki prediction models.
AlgorithmDescriptor Q 2 R M S E C V M A E C V R 2 R M S E T M A E T
SVMDaylight0.725 ± 0.0120.408 ± 0.0090.317 ± 0.0050.7660.4190.320
E-state0.502 ± 0.0100.550 ± 0.0050.417 ± 0.0060.5360.5900.448
ECFP40.744 ± 0.0080.394 ± 0.0060.318 ± 0.0040.7610.4240.325
MACCS0.684 ± 0.0060.438 ± 0.0040.344 ± 0.0040.6870.4850.362
BaggingDaylight0.742 ± 0.0180.395 ± 0.0130.307 ± 0.0090.7790.4080.312
E-state0.677 ± 0.0180.442 ± 0.0120.348 ± 0.0080.6420.5190.393
ECFP40.778 ± 0.0120.367 ± 0.0100.291 ± 0.0080.7800.4070.305
MACCS0.697 ± 0.0240.428 ± 0.0160.334 ± 0.0130.7500.4330.323
GBDTDaylight0.723 ± 0.0100.410 ± 0.0070.326 ± 0.0050.7550.4290.332
E-state0.671 ± 0.0130.447 ± 0.0090.356 ± 0.0070.6230.5320.410
ECFP40.759 ± 0.0070.382 ± 0.0050.309 ± 0.0040.7570.4270.329
MACCS0.686 ± 0.0110.437 ± 0.0080.340 ± 0.0070.7030.4720.371
XGBoostDaylight0.723 ± 0.0220.410 ± 0.0150.317 ± 0.0110.7660.4190.316
E-state0.683 ± 0.0320.438 ± 0.0200.342 ± 0.0140.6480.5140.385
ECFP40.771 ± 0.0140.373 ± 0.0110.301 ± 0.0090.8160.3710.292
MACCS0.696 ± 0.0200.429 ± 0.0130.337 ± 0.0110.7450.4370.330
Table 3. The results of internal and external validation of IC50 prediction models.
Table 3. The results of internal and external validation of IC50 prediction models.
AlgorithmDescriptor Q 2 R M S E C V M A E C V R 2 R M S E T M A E T
SVMDaylight0.726 ± 0.0060.424 ± 0.0050.338 ± 0.0040.7440.4430.353
E-state0.487 ± 0.0080.580 ± 0.0040.455 ± 0.0030.5450.5900.455
ECFP40.759 ± 0.0060.398 ± 0.0050.318 ± 0.0040.7630.4260.342
MACCS0.639 ± 0.0050.487 ± 0.0040.381 ± 0.0030.6820.4940.391
BaggingDaylight0.719 ± 0.0160.429 ± 0.0120.343 ± 0.0080.7120.4690.366
E-state0.642 ± 0.0200.485 ± 0.0130.376 ± 0.0100.6280.5340.426
ECFP40.757 ± 0.0150.399 ± 0.0120.318 ± 0.0080.7220.4620.362
MACCS0.674 ± 0.0170.462 ± 0.0110.364 ± 0.0080.6810.4940.396
GBDTDaylight0.685 ± 0.0070.455 ± 0.0050.368 ± 0.0030.7060.4750.378
E-state0.555 ± 0.0060.540 ± 0.0030.428 ± 0.0020.5840.5640.449
ECFP40.673 ± 0.0040.463 ± 0.0030.374 ± 0.0030.7030.4770.386
MACCS0.579 ± 0.0050.525 ± 0.0030.418 ± 0.0030.6100.5460.437
XGBoostDaylight0.742 ± 0.0200.411 ± 0.0150.325 ± 0.0110.7460.4410.347
E-state0.660 ± 0.0220.472 ± 0.0140.368 ± 0.0110.6640.5070.389
ECFP40.806 ± 0.0130.357 ± 0.0110.290 ± 0.0070.7840.4070.328
MACCS0.699 ± 0.0200.444 ± 0.0140.349 ± 0.0090.7270.4570.367
Table 4. The results of internal and external validation of EC50 prediction models.
Table 4. The results of internal and external validation of EC50 prediction models.
AlgorithmDescriptor Q 2 R M S E C V M A E C V R 2 R M S E T M A E T
SVMDaylight0.784 ± 0.0090.505 ± 0.0100.409 ± 0.0080.8090.5320.420
E-state0.665 ± 0.0130.629 ± 0.0110.509 ± 0.0120.7160.6490.492
ECFP40.772 ± 0.0080.518 ± 0.0090.416 ± 0.0060.8440.4810.382
MACCS0.758 ± 0.0100.534 ± 0.0110.423 ± 0.0110.7510.6070.488
BaggingDaylight0.765 ± 0.0150.527 ± 0.0160.415 ± 0.0130.7180.6470.492
E-state0.725 ± 0.0220.570 ± 0.0220.454 ± 0.0150.7350.6260.474
ECFP40.782 ± 0.0170.507 ± 0.0180.400 ± 0.0160.8440.4800.367
MACCS0.746 ± 0.0250.547 ± 0.0250.431 ± 0.0200.7660.5890.450
GBDTDaylight0.772 ± 0.0140.518 ± 0.0150.408 ± 0.0130.7450.6140.465
E-state0.731 ± 0.0190.563 ± 0.0190.458 ± 0.0120.7770.5750.428
ECFP40.775 ± 0.0120.515 ± 0.0130.402 ± 0.0110.8320.4990.404
MACCS0.742 ± 0.0120.552 ± 0.0120.432 ± 0.0120.7590.5970.475
XGBoostDaylight0.771 ± 0.0300.519 ± 0.0300.409 ± 0.0230.7770.5750.443
E-state0.729 ± 0.0260.566 ± 0.0250.439 ± 0.0220.7720.5810.445
ECFP40.778 ± 0.0210.512 ± 0.0220.395 ± 0.0170.8400.4870.380
MACCS0.751 ± 0.0190.542 ± 0.0190.422 ± 0.0160.6990.6680.501
Table 5. The results of internal and external validation of prediction models and previous studies.
Table 5. The results of internal and external validation of prediction models and previous studies.
Ki ModelIC50 ModelEC50 ModelKristam et al. [12]Wang et al. [14]
Q 2 0.7780.8060.7840.90.522
R 2 0.7800.7840.8090.750.839
n661189436762236
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wei, X.; Huang, T.; Yang, Z.; Pan, L.; Wang, L.; Ding, J. Quantitative Predictive Studies of Multiple Biological Activities of TRPV1 Modulators. Molecules 2024, 29, 295. https://doi.org/10.3390/molecules29020295

AMA Style

Wei X, Huang T, Yang Z, Pan L, Wang L, Ding J. Quantitative Predictive Studies of Multiple Biological Activities of TRPV1 Modulators. Molecules. 2024; 29(2):295. https://doi.org/10.3390/molecules29020295

Chicago/Turabian Style

Wei, Xinmiao, Tengxin Huang, Zhijiang Yang, Li Pan, Liangliang Wang, and Junjie Ding. 2024. "Quantitative Predictive Studies of Multiple Biological Activities of TRPV1 Modulators" Molecules 29, no. 2: 295. https://doi.org/10.3390/molecules29020295

Article Metrics

Back to TopTop