Next Article in Journal
Molecular Regulation of Heme Oxygenase-1 Expression by E2F Transcription Factor 2 in Lung Fibroblast Cells: Relevance to Idiopathic Pulmonary Fibrosis
Next Article in Special Issue
Mechanistic Insights into the Protection Effect of Argonaute–RNA Complex on the HCV Genome
Previous Article in Journal
Advances in Molecular Biomarkers in Cardiology
Previous Article in Special Issue
Drug-Disease Association Prediction Using Heterogeneous Networks for Computational Drug Repositioning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EP-Pred: A Machine Learning Tool for Bioprospecting Promiscuous Ester Hydrolases

1
Department of Life Sciences, Barcelona Supercomputing Center (BSC), 08034 Barcelona, Spain
2
Department of Applied Biocatalysis, ICP, CSIC, 28049 Madrid, Spain
3
Catalan Institution for Research and Advanced Studies (ICREA), 08010 Barcelona, Spain
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Biomolecules 2022, 12(10), 1529; https://doi.org/10.3390/biom12101529
Submission received: 2 September 2022 / Revised: 11 October 2022 / Accepted: 18 October 2022 / Published: 21 October 2022

Abstract

:
When bioprospecting for novel industrial enzymes, substrate promiscuity is a desirable property that increases the reusability of the enzyme. Among industrial enzymes, ester hydrolases have great relevance for which the demand has not ceased to increase. However, the search for new substrate promiscuous ester hydrolases is not trivial since the mechanism behind this property is greatly influenced by the active site’s structural and physicochemical characteristics. These characteristics must be computed from the 3D structure, which is rarely available and expensive to measure, hence the need for a method that can predict promiscuity from sequence alone. Here we report such a method called EP-pred, an ensemble binary classifier, that combines three machine learning algorithms: SVM, KNN, and a Linear model. EP-pred has been evaluated against the Lipase Engineering Database together with a hidden Markov approach leading to a final set of ten sequences predicted to encode promiscuous esterases. Experimental results confirmed the validity of our method since all ten proteins were found to exhibit a broad substrate ambiguity.

1. Introduction

Enzymes are of great interest for a vast majority of industries, partially due to the increasing concerns over environmental issues. Among the many classes of enzymes, hydrolases stand out for their industrial relevance because of their high stereoselectivity, commercial availability and stability in organic solvents [1]. Indeed, the demand for newer and better hydrolases that can work in industrial settings has only increased exponentially over the years. Specifically, ester hydrolases (EC 3.1), which hydrolyze ester bonds, have received considerable attention, and are extensively used in various areas such as food, detergents, agriculture, pharmaceuticals, and so on [2].
Searching for new esterase candidates is not trivial, in fact, there are strict requirements regarding stability, activity and substrate promiscuity which are difficult to find conjointly in natural enzymes [3]. Actually, one of the most common issues that an industrial enzyme will face is having low substrate promiscuity [4]; the ability to catalyze a specific reaction for a variety of different substrates. It is a desirable characteristic since one single enzyme could be used for multiple applications, thus reducing the cost and time of development and production of multiple biocatalysts [5].
While some substrate promiscuous enzymes might suffer from limited stereoselectivity and lower catalytic rates [6,7], they are typically enzymes prone to accept enzyme engineering, which could restore these properties. In previous studies, we have investigated the determinants of substrate ambiguity for esterases at a molecular level, establishing rules for its prediction [4], and introducing a significant increase in the number of substrates hydrolyzed through engineering [8]. However, it must be noted that the necessary molecular metrics must be computed from the 3D structures which are rarely available, and that they involve significant computational time. Even with the recent advancements in the accuracy of deep learning structural predictions, such as AlphaFold 2.0 [9], it is still unfeasible to analyze substrate specificity from the ever growing number of annotated sequences. For instance, as of 9 March 2021, the Lipase Engineering Database (LED) [10], which holds data on esterases/lipases and a few other homologous sequences, contains about 280,638 entries. In addition, AlphaFold 2.0 models tend to generate APO structures, with a significant modification (mostly volume reduction) in the active site that precludes, for example, efficient ligand rigid docking [11]. These changes would largely affect promiscuity predictions using the molecular descriptors.
Therefore, methods to directly identify substrate promiscuity from sequence alone would greatly increase the efficiency of bioprospecting for new esterase candidates, a task ideal for machine learning algorithms (see for example, other recently developed methods for activity predictions [12]). Several studies have already predicted enzyme substrate promiscuity using molecular descriptors [13] or machine learning [14,15,16] approaches, although for other enzyme families. In addition, there are several differences in our approach. First, other studies mainly used training samples listed in databases such as the Kyoto Encyclopedia of Genes and Genomes (KEGG), which includes reaction data for most of the studied enzyme families. In these databases the number of tested compounds is usually not large, hindering the correct identification of promiscuous enzymes. Additionally, those studies that revolved around single enzyme families are usually constrained by the number of samples from which to derive the molecular descriptors or the classification models which might decrease their applicability. Second, the goal of the developed methods seems to differ. Previous projects evaluate whether a specific compound will be catalyzed by a set of characterized enzymes or the reverse to find novel substrates or enzymes. Our approach, in addition, tries to classify how promiscuous an enzyme might be in a bioprospecting setting.
Here we report the development and application of an ensemble binary classifier trained on a dataset of 145 diverse esterases and 96 substrates [4], using a combination of physicochemical and evolutionary feature vectors extracted from the primary sequence. Our classifier, named EP-pred, combines three types of classification algorithms: SVM (support vector machines), KNN (k-nearest neighbors) and RidgeClassifier, one of the linear models implemented in Scikit-Learn. EP-pred was then evaluated against LED and from those predicted to be positives, a final set of ten sequences were isolated and tested experimentally. All selected enzymes were confirmed to be substrate promiscuous, highlighting the potential of our machine learning bioprospecting method.

2. Materials and Methods

2.1. Esterase Dataset

The dataset employed to train the models is the same used in our previous molecular modeling studies and it is formed by 145 diverse microbial ester hydrolases with pairwise sequence identities ranging from 0.2 to 99.7% and an average pairwise identity of 13.7% [4]. The heterogeneity of the sequences can be attributed to the diversity of the source from which they were isolated, including both terrestrial bacteria from 28 geographically distinct sites and marine bacteria. The phylogenetic analysis performed in the previous study further supports the diversity of the source bacteria since they were found to be distributed across the phylogenetic tree.
The substrate profiles of the enzymes were assessed on a set of 96 diverse esters with the most promiscuous one capable of hydrolyzing 72 esters out of 96 tested and the least promiscuous one only capable of catalyzing 1 out of 96 substrates. The distinction between promiscuous and not promiscuous, or positive and negative classes, was also established according to the threshold of the previous study: 20 substrates.

2.2. Lipase Engineering and Uniref 50 Databases

The lipase engineering database (https://led.biocatnet.de/, accessed on 9 March 2021), used to evaluate the final model, gathers information on the sequence, structure and function of esterases/lipases and other related proteins sharing the same a/b hydrolase fold. The whole sequence database was downloaded in Fasta format containing 280,638 entries.
The evolutionary-related features were based on PSSM (Position Specific Scoring Matrix) profiles which were generated with Psi-Blast by querying the input sequence against the Uniref 50. Uniref or The UniProt Reference Clusters contains records of the Uniprot Knowledgebase and the Uniparc sequence archive at several resolutions, 100%, 90% and 50%, each one generated from the clustering of the previous one. Thus, Uniref50 is generated from clustering the UniRef90 seed sequences at the identity threshold of 50% [17].

2.3. Feature Extraction

Two web servers Possum [18] and iFeature [19] were used to extract evolutionary information and physicochemical properties, respectively, from all protein sequences. iFeature can generate 53 different types of descriptors, from which 32 were extracted using the default parameters resulting in a total of 2274 feature vector dimensions. The rest of the feature types were discarded because they can only be applied to sequences of the same length. These features might be simplistic descriptors like the amino acid composition (AAC), which counts the frequency of each amino acid in the sequence. However, there are also more elaborate forms of descriptors that account for the distribution, transition, or correlation of different properties, like hydrophobicity, along the sequence.
Possum generates features based on the PSSM (Position Specific Scoring Matrix) profiles that contain evolutionary information of the sequences since it specifies the scores for observing particular amino acids at specific positions of the sequence. Although very informative, the downside of these profiles is that they depend on the length of the sequences which hampers their direct use as features for machine learning applications. By applying different matrix transformations to make them length-independent, Possum was able to generate 18 different descriptors that were extracted resulting in a vector of 18,730 dimensions. Some transformations are inspired by sequence-based features, such as ACC which reduces the PSSM profile from a matrix of L × 20 dimensions, L being the length of a sequence, to a vector of 20 dimensions by averaging the scores of the rows in the PSSM. While other transformations are more complex constructs that first scale, filter, or group the values in the PSSM and then apply various operations to fix the dimensions of the matrix.
The concatenation of both feature vectors yields a feature set of 21,000 dimensions for the esterase dataset. After generating the features, some cleaning was needed because many columns had zeros or identical values in most of the rows which carried little information. As a result, 2274 iFeature and 18,730 Possum features were reduced to 1203 and 14,606 dimensions, respectively.

2.4. Feature Selection

Even with the cleaning, the number of original features remained exceedingly high, therefore feature selection was needed to eliminate noise and avoid overfitting. As recommended for this step [20], data was split into two sets, a test set and a training set and the selection was performed on the training set only. In addition, the number of dimensions was reduced to less than ½ of the number of samples as it was shown to reduce overfitting.
It must be noted that the dimensionality reduction of iFeature and Possum descriptors was carried out independently and concatenated later to generate the features. However, the proportion of the two descriptors was not even because evolutionary-related features seem to bear more information [18,21,22], so they were given a larger weight during the construction of the feature set compared to the iFeature descriptors. Furthermore, following this idea, five other sets of features with varying dimensions were also constructed and tested. The whole process was repeated ten times, one for each of the selection algorithms resulting in a total of 60 feature sets.
The selection methods could be divided into three categories: filter methods that assessed the degree of dependence between the features and the labels, wrapper and embedded methods that applied machine learning algorithms to rank those features based on their relevance for the performance [23].
A total of 5 libraries were used to implement the methods from the different categories: (I) ITMO_FS [24] which provided filter methods such as Chi-square and Information gain. (II) Boruta [25] a library containing a single embedded method. (III) Scikit-feature [26] that implemented the filter methods MRMR (minimum redundancy maximum relevancy) and CIFE (conditional infomax feature extraction). (IV) Scikit-learn that provided the filter methods mutual information and fisher score; the wrapper method RFE (recursive feature elimination) combined with a linear model or SVM and the embedded method random forest. Finally, (V) XGBoost [27], like boruta, is a library that contains only an embedded method.

2.5. Model Training

SVM, KNN and RidgeClassifier, which are all implemented in Scikit-Learn, were selected for classification. To correctly evaluate the model’s performance, we employed a similar strategy to nested cross-validation [20]. The data was split into a 20% test set and 80% training set 5 times, each time generating different sets. Then, for each split, the training set was used for model development to find the optimal sets of hyperparameters using 5-fold cross-validation while the test set was used for the evaluation. It generates five measurements from models with different sets of hyperparameters that can be used to compute the statistics on the model’s performance.

2.6. Performance Metrics

Using the TP (true positive), TN (true negative), FN (false negative) and FP (false positive) values, the precision (Pr), recall (Re), F1 [28] and Matthew’s correlation coefficient (MCC) were calculated [29] to evaluate the performance of the models.
Pr = TP FP + TP
Re = TP FN + TP  
F 1 = 2   Pr Re Pr +   Re
MCC = TP   TN FP FN TP + FP TP + FN TN + FP TN + FN

2.7. Applicability Domain

There are several aspects that might affect the reliability of the model’s predictions apart from the performance metrics. Indeed, there should be limitations in the applicability of the models to be used only on those samples that are similar to the training samples, because otherwise, it would be predicting sequences that it has not seen and fitted before. In other words, we should define the applicability domain (AD) of the models and filter the predictions accordingly.
There are several approaches to define the similarity or the AD, all within the feature or descriptor space, but we decided to follow one inspired by KNN [30]. In this approach, a distance threshold was computed for each training sample and compared to the Euclidean distance between a new sample and each training sample. If any of the distances between the new and the training samples was less or equal to the threshold associated with that training sample, the prediction was deemed reliable and kept.

2.8. Hidden Markov Model (HMM) Profiles

LED contains sequences other than esterases/lipases so it would be wise to filter them and keep only those most likely to be esterases before the predictions. For this purpose, we employed the HMM profiles, probabilistic models that capture the evolutionarily conserved patterns revealed by multiple sequence alignments (MSA). They allow more sensitive homology searches than blast while retaining the speed.
We followed the protocol described by Pérez-García et al [31]. The program used to build such profiles was HMMER (http://hmmer.org/, accessed on 10 March 2021), using a MSA of the esterases with 35 or more substrates. The MSA was generated by T-Coffee with the default parameters. The program can then use the HMM profile to search for homologs in sequence databases and filter them based on E-values; here we used an E-value cutoff of 10−10 to add precision. Notice that the resulting HMM model is aimed at filtering esterases rather than promiscuity. In fact, when applied to the training dataset, it cannot distinguish well between promiscuous and non-promiscuous enzymes, with a precision score of 0.6 at E-value of 0.001.

2.9. Homology Modelling (HM) and Active Site Analysis

The top selected predictions were modeled using ModWeb [32], a web server for protein structure modeling, which automatically generates a homology model of the target sequence. Note that here we only aimed at a fast structural method to generate approximate active site structures so we could discard clearly wrong ones. Notice that for those sequences tested experimentally, we also constructed the Alphafold2 models (not yet available at the conception of the project). AlphaFold2 results clearly agreed with the ones predicted by ModWeb, with low values of RMSD when compared (see, Table S6). We also checked the structural quality of the models from the top 10 promiscuous esterases with ProSA-web [33]. It compares the structure models with experimentally determined proteins from Protein Data Bank and estimates a Z-score for each model, the lower the better.
The active site of the homology models was analyzed to filter out esterases with the catalytic triads not arranged in an active conformation to make sure they were indeed esterases. Next, the properties of their active site were calculated using SiteMap, Schrodinger [34,35] which includes hydrophobicity, enclosure and exposure that gave an idea of how solvent-exposed the cavity of the enzymes was. These metrics are relevant because we found out [8] that the active site of the promiscuous esterases share some common physicochemical features such as high hydrophobicity, larger volumes and are more enclosed compared to their non-promiscuous counterparts. Accordingly, we used this information to rank and isolate the 10 sequences tested in this paper when we needed to reduce the number of candidates.

2.10. Enzyme Source, Production, and Purification

The sequences encoding AJP48854.1, ART39858.1, PHR82761.1, WP_014900537.1, WP_026140314.1, WP_042877612.1, WP_059541090.1, WP_069226497.1, WP_089515094.1 and WP_101198885.1 were used as templates for gene synthesis (GenScript Biotech, EG Rijswijk, The Netherlands), and genes were codon-optimized to maximize expression in Escherichia coli. Genes were flanked by BamHI and HindIII (stop codon) restriction sites and inserted in a pET-45b(+) expression vector with an ampicillin selection marker (GenScript Biotech, Rijswijk, The Netherlands), which was further introduced into E. coli BL21(DE3).
The soluble N-terminal histidine (His) tagged proteins were produced and purified (98% purity, as determined by SDS–PAGE analysis using a Mini PROTEAN electrophoresis system, Bio-Rad, Madrid, Spain) at 4 °C after binding to a Ni-NTA His-Bind resin (Merck Life Science S.L.U., Madrid, Spain), as previously described [4,7], and stored at −86 °C until use at a concentration of 10 mg mL−1 in 40 mM 4-(2-hydroxyethyl)-1-piperazineethanesulfonic acid (HEPES) buffer (pH 7.0).

2.11. Activity Tests

The hydrolysis of esters was assayed using a pH indicator assay in 384-well plates (ref. 781162, Greiner Bio-One GmbH, Kremsmünster, Austria) at 40 °C and pH 8.0 in a Synergy HT Multi-Mode Microplate Reader in continuous mode at 550 nm over 24 h (extinction coefficient (ε) of phenol red, 8450 M−1 cm−1), as reported [36,37]. The conditions for determining the specific activity (units mg−1) were as follows: [proteins]: 270 μg mL−1; [ester]: 20 mM; reaction volume: 44 μL; T: 30 °C; and pH: 8.0 (5 mM 4-(2-hydroxyethyl)-1-piperazinepropanesulfonic acid (EPPS) buffer). In all cases, all values in triplicate were corrected for nonenzymatic transformation, with the absence of activity defined as having at least a twofold background signal. In all cases, the activity was calculated by determining the absorbance per minute from the slopes generated [38].
The activity toward the model esters p-nitrophenyl (p-NP) acetate (ref. N-8130; Merck Life Science S.L.U., Madrid, Spain), propionate (Santa Cruz Biotechnology, Inc., Heidelberg, Germany, ref. sc-256813) and butyrate (ref. N-9876; Merck Life Science S.L.U., Madrid, Spain) was assessed in 5 mM EPPS buffer at pH 8.0 and 30 °C by monitoring the production of 4-nitrophenol at 348 nm (pH-independent isosbestic point, ε = 4147 M−1 cm−1) over 5 min and determining the absorbance per minute from the generated slopes [36]. The reactions were performed in 96-well plates (ref. 655801, Greiner Bio-One GmbH, Kremsmünster, Austria). For specific activity determinations, the following conditions were used: [proteins]: 7 μg mL−1; [ester]: 1 mM; reaction volume: 200 μL; T: 30 °C; and pH: 8.0 (5 mM EPPS buffer). For Km and kcat determinations (using p-NP propionate), the following conditions were used: [proteins]: 0.06–25 μg mL−1; [ester]: 0–0.04 mM; reaction volume: 100 μL; T: 30 °C; and pH: 8.0 (5 mM EPPS buffer). The values correspond to the fit obtained from the regression of the data (each obtained in triplicates) using SigmaPlot 14.0 software. kcat was calculated by using the following equation: kcat = Vmax/[E], where [E] = total enzyme (for raw data see Supplementary Material).
Meta-cleavage product (MCP) hydrolase activity was assayed using 2-hydroxy-6-oxo-6-phenylhexa-2,4-dienoate (HOPHD) and 2-hydroxy-6-oxohepta-2,4-dienoate (HOHD), freshly produced as described [38]. The reactions were performed at 30 °C in 96-well plates (ref. 655801, Greiner Bio-One GmbH, Kremsmünster, Austria), and they contained 7.0 μg mL−1 proteins and 0.2 mM HOPHD or HODH in a total volume of 200 μL 50 mM K/Na-phosphate (pH 7.5) buffer (this buffer was shown to be optimal for measuring MCP hydrolytic activity [38]. Hydrolysis was monitored at 388 nm (for HOPHD) or 434 nm (for HOHD) over 5 min and the absorbance per minute was determined from the generated slopes [38].

3. Results and Discussion

3.1. Model Buildup

The accuracy of machine learning classifiers depends greatly on the features and the hyperparameters used. To construct the features, we derived physicochemical and evolutionary information from the sequences via two webservers iFeatures [19] and Possum [18] (see Tables S1 and S2, respectively), reduced their dimension through feature selection and built a total of 60 sets of features to be tested.
The classifiers were then trained on one of the feature sets during which the hyperparameters were tuned using 5-fold cross-validation (CV). Lastly, the model with the optimal hyperparameters was evaluated on the test set. The process was repeated five times, one for each of the data splits, from which the statistics on the model performance were computed and compared against other models using a distinct feature set (Figure 1). Among the 60 features, two sets generated the best models, called hereafter ch_20 and random_30, since they were produced by Chi-squared and Random Forest feature selection methods, respectively. Both SVM and RidgeClassifier performed the best when trained on ch_20 and both showed a mean MCC score of 0.54 for the training set and a mean MCC score of around 0.62 for the test set. In contrast, KNN performed the best when trained on random_30 and showed a mean MCC score of 0.67 for the training set and a mean MCC score of 0.65 for the test set (Figure 2). KNN slightly outperformed the others in the training set score which might imply that the algorithm might be more suited at fitting to this type of data.
PSSM-based features, which encode the evolutionary information of the sequences, seem to be more relevant for the prediction accuracy. In both selected sets, PSSM features proportion was larger than physicochemical ones. For instance, 19 out of 25 descriptors in random_30 were extracted from PSSM. This is in line with our previous findings [4] where phylogeny was a predictive marker of the substrate promiscuity in ester hydrolases.
The MCC scores for the three classifiers, which indicate the correlation between the predicted and the true labels, are good, but different models can be grouped to improve the performance since they have different biases that might complement each other. There are many possible combinations because each machine learning algorithm was trained on five different data sets resulting in five distinct classifiers. However, to generate the combinations, only two to three models from each algorithm were chosen, those with better MCC scores. EP-pred, which aggregates all the models, two SVM, three RidgeClassier and two KNN models, displayed a mean MCC score of 0.73 for the training set and a score of 0.72 for the test set (Figure 2). The observed increase in the models’ scores can be attributed to the fact that only samples for which the predictions between different classifiers agreed were kept for scoring, thus making the predictions more robust.

3.2. The Workflow for In Silico Bioprospecting

LED gathers sequences from various families apart from esterases/lipases, which is why we applied an HMM profile, built from the esterase dataset, as a filtering step and ended up with approximately 70,000 sequences. Then, the final model EP-pred was evaluated against them and predicted around 500 positive (promiscuous) sequences which were still too much for the experimental validation. Thus, several filters were applied to decrease the number of hits to a final set of ten.
The top 100 sequences according to E-values returned by HMM were selected to be modeled and their active site cavity analyzed in search of the catalytic triad and geometric descriptors. Only 73 sequences passed this second filter and were forwarded to the subsequent analysis by SiteMap, a widely used binding site analysis tool, which then generated various binding cavity descriptors. As seen in our previous engineering studies, two metrics: hydrophobicity, and the ratio of enclosure/exposure, were useful in ranking promiscuity, see Table S3; thus, we used these to rank the final set of ten proteins for experimental validation picking those that intersected at the top in both metrics (Figure 3).

3.3. The Experimental Validation

All ten recombinant presumptive substrate promiscuous hydrolases (AJP48854.1, ART39858.1, PHR82761.1, WP_014900537.1, WP_026140314.1, WP_042877612.1, WP_059541090.1, WP_069226497.1, WP_089515094.1 and WP_101198885.1) were successfully expressed in soluble form and purified by nickel affinity chromatography. Then, three model p-nitrophenyl (p-NP) ester substrates with different chain lengths: p-NP acetate (C2), p-NP propionate (C3), and p-NP butyrate (C4) were first used to determine the substrate specificity of the enzymes. Their hydrolytic activity was assessed and recorded under standard assay conditions described in Section 2.11. We found specific activities ranging from 5.85 U mg−1 to 2.19 U mg−1 for p-NP propionate, which was the best substrate in all cases (Table 1).
Once the esterase activity was confirmed, we further tested the hydrolytic activity towards a set of 96 structurally different esters based on Tanimoto-Combo similarity [4,39]. As shown in Figure 4, all enzymes were able to hydrolyze an ample set of esters, ranging from 27 (for AJP48854.1) to 68 (for WP_069226497.1). The specific activity ranges from 6. 50 (WP_069226497.1, being the most active) to 0.01 U mg−1 (WP_014900537.1, being the least active), depending on the substrate (Table S4). According to the criteria previously established [4], nine of the enzymes could be considered as having high-to-prominent substrate promiscuity as they hydrolyze 30 or more esters, whereas one (AJP48854.1) could be considered as moderately substrate promiscuous as it used less than 30 esters but more than 10 (a number below which an esterase could be considered substrate specific). Based only on the number of esters converted (Figure 5), these enzymes could be ranked among the hydrolases with the highest substrate promiscuity within a total of 145 esterases previously tested with a similar set of esters. Kinetic characterization using the model ester substrate p-NP propionate confirmed the high affinity of nine out the ten tested hydrolases for this substrate (Km from14.5 to 48.7 µM) and the high conversion rates (kcat from 2060 to 5043 min−1) of six of them (Table 2; Figure S1).
A database search indicated that all ten hydrolases showed from 60 to 82.9% identity with the meta-cleavage product hydrolase (MCP hydrolase) from Pseudomonas fluorescens IP01 (CumD) [40]. These hydrolases participate in the aerobic pathways for the bacterial degradation of aromatic carbons, in which aromatic compounds are cleaved into meta-ring fission compounds [35,38].
To check whether the ten proteins herein retrieved do show such activity, we tested two common meta-cleavage compounds 2-hydroxy-6-oxo-6-phenylhexa-2,4-dienoate (HOPHD) and 2-hydroxy-6-oxohepta-2,4-dienoate (HOHD) [38,40]. Using the standard assay conditions described in Section 2.11, we found all ten hydrolases converted HOPHD and HOHD, with HOHD being the preferred (from 132- to 273-fold) substrate for eight of the hydrolases (AJP48854.1, WP_042877612.1, WP_059541090.1, ART39858.1, WP_089515094.1, WP_026140314.1, PHR82761.1 and WP_014900537.1), and HOPHD the preferred (~2.5-fold) for two of them (WP_101198885.1 and WP_069226497.1) (Figure 4). The specific activity ranged from 23.2 to 0.4 U mg−1 for HOHD and from 89.04 to 0.8.6 U mg−1 for HOPHD, which indicates that both were the preferred substrates among all substrates tested (including non-activated and synthetic, activated p-NP esters).
When we compared the ten sequences to the training set, the average sequence identity dropped to 17%, with the maximum sequence identity being 67% and the minimum being 2% (Table S5). Additionally, if we compared the predicted sequences to the top 15 most promiscuous enzymes in the training set, which includes CalB, the average sequence identity dropped even further towards 15% with the maximum identity being 19% and the minimum identity being around 7%. So even if they have matching promiscuity their sequences are very different. In fact, sequence identity between those top 15 esterases is relatively low, with an average identity of 29%, a maximum identity of 51% and a minimum identity of 9%. Thus, as described before [4], sequence identity cannot be considered a predictor of promiscuity.
At the structural level, we did not find major differences when comparing the ten predicted sequences against the promiscuous or non-promiscuous training set, with a mean RMSD of 3.11 Å with the top 15 promiscuous esterases and 3.08 Å for the remaining of the training set (see, Table S7). Still, all ten selected enzymes share common active site physicochemical properties with the most promiscuous enzymes, derived from our filtering of the properties performed with SiteMap.

4. Conclusions

An essential aspect of implementing enzymes into industrial processes is the ability to perform fast and accurate bioprospecting. Modern genomic techniques are providing us with millions of new un-annotated sequences that have the potential of becoming new biocatalysts. However, techniques capable of working at a sequence level are necessary to cherry-pick those enzymes with favorable industrial qualities (or particular uses) in a feasible time. Here, ML methods seem an appropriate choice. As in any ML application, one of the main prerequisites is having enough (and diverse) data to train the model. We show in this study that if enough data is available, in our case a cross activity map between 145 enzymes and 96 substrates [4], a ML model can learn to distinguish those sequences encoding substrate promiscuity in ester hydrolases.
The experimental results herein provided, clearly demonstrate the validity of our ensemble classifier EP-pred in the prediction of substrate promiscuity of ester hydrolases. More importantly, the hydrolases herein retrieved were capable of hydrolyzing C-O bonds in an ample set of esters, but also C-C bonds in meta-cleavage products of catechol and biphenyl derivatives. Based on the sequence similarity and the substrate specificity and preferences, the enzymes herein retrieved could be classified as promiscuous MCP hydrolases with the ability to also convert a broad range of esters. This demonstrates the ability of the EP-pred system to identify hydrolases with the ability to catalyze ester hydrolysis for a broad range of different substrates even if it might not be the main reaction catalyzed by the enzymes.
The overall workflow applied in this project is not complex: (I) Filtering of sequence databases, such as a metagenomic database using HMM profiles to isolate ester hydrolases, (II) Prediction of substrate promiscuous ester hydrolases with EP-pred and (III) Homology modeling and structural analysis of the binding site cavity with SiteMap scores. Notice, that the workflow still uses molecular predictors to rank the final list of 100 sequences provided by the ML model; the sequence predictor is used as a means for selecting a short list of candidates, for which the structure generation and molecular descriptors extraction is an easy task. This was forced mainly by the academic (low budget) nature of our research. In a realistic industrial setting, hundreds of enzymes could be expressed and tested in vitro, thus possibly bypassing the last structural characterization step. Such an ML-only procedure would also facilitate the implementation of a similar approach for other enzymatic properties, as far as a comprehensive data training set is available; future studies will determine the wide applicability of this approach.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/biom12101529/s1, Figure S1: Calculation of kinetic parameters for p-NP propionate; Table S1: The features extracted from iFeature; Table S2: The features extracted from possum; Table S3: The sitemap scores of the 73 sequences that passed the first filter; Table S4: The activity assay of the ten enzymes versus the 96 ester substrates; Table S5: Pairwise sequence identity between the training set and the predicted sequences; Table S6: Quality of the models from ProsA-Web; Table S7: The RMSD of the ten sequences against the training set.

Author Contributions

Conceptualization, R.X. and V.G.; methodology, R.X., A.R.-M., L.F.-L., M.F. and V.G.; software, R.X. and A.R.-M.; validation, L.F.-L. and M.F.; formal analysis, R.X., L.F.-L., A.R.-M., M.F. and V.G.; investigation, R.X. and L.F.-L.; writing—review and editing, R.X., A.R.-M., L.F.-L., M.F. and V.G.; supervision, V.G. and M.F.; funding acquisition, M.F. and V.G. All authors have read and agreed to the published version of the manuscript.

Funding

This study was conducted under the auspices of the FuturEnzyme and Oxipro Projects funded by the European Union’s Horizon 2020 Research and Innovation Programme under Grant Agreement No. 101000327 and 101000607. We also acknowledge financial support under Grants PID2020-112758RB-I00 (M.F.), PID2019-106370RB-I00 (V.G.) and PDC2021-121534-I00 (M.F.) and PID2019-106370RB-I00/AEI/10.13039/501100011033 (A.R.-M.), from the Ministerio de Ciencia e Innovación, Agencia Estatal de Investigación (AEI) (Digital Object Identifier 10.13039/501100011033), Fondo Europeo de Desarrollo Regional (FEDER) and the European Union (“NextGen-erationEU/PRTR”), and Grant 2020AEP061 (M.F.) from the Agencia Estatal CSIC.

Data Availability Statement

The code, features and the results of the training can be downloaded from https://github.com/etiur/EP-pred.

Acknowledgments

The authors acknowledge Rafael Bargiela for supporting the preparation of Figure 4, and David Almendral for the enzyme purification.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alcántara, A.; Hernaiz, M.-J.; Sinisterra, J.-V. 3.28 Biocatalyzed Production of Fine Chemicals. Compr. Biotechnol. 2017, 3, 334–373. [Google Scholar] [CrossRef]
  2. Panda, T.; Gowrishankar, B.S. Production and applications of esterases. Appl. Microbiol. Biotechnol. 2005, 67, 160–169. [Google Scholar] [CrossRef] [PubMed]
  3. Kamble, A.; Srinivasan, S.; Singh, H. In-Silico Bioprospecting: Finding Better Enzymes. Mol. Biotechnol. 2019, 61, 53–59. [Google Scholar] [CrossRef] [PubMed]
  4. Martínez-Martínez, M.; Coscolín, C.; Santiago, G.; Chow, J.; Stogios, P.J.; Bargiela, R.; Gertler, C.; Navarro-Fernández, J.; Bollinger, A.; Thies, S.; et al. Determinants and Prediction of Esterase Substrate Promiscuity Patterns. ACS Chem. Biol. 2018, 13, 225–234. [Google Scholar] [CrossRef] [Green Version]
  5. Schmid, A.; Dordick, J.S.; Hauer, B.; Kiener, A.; Wubbolts, M.; Witholt, B. Industrial biocatalysis today and tomorrow. Nature 2001, 409, 258–268. [Google Scholar] [CrossRef]
  6. Braakman, R.; Smith, E. Metabolic evolution of a deep-branching hyperthermophilic chemoautotrophic bacterium. PLoS ONE 2014, 9, e87950. [Google Scholar] [CrossRef] [Green Version]
  7. Giunta, C.I.; Cea-Rama, I.; Alonso, S.; Briand, M.L.; Bargiela, R.; Coscolín, C.; Corvini, P.F.-X.; Ferrer, M.; Sanz-Aparicio, J.; Shahgaldian, P. Tuning the Properties of Natural Promiscuous Enzymes by Engineering Their Nano-environment. ACS Nano 2020, 14, 17652–17664. [Google Scholar] [CrossRef]
  8. Roda, S.; Fernandez-Lopez, L.; Cañadas, R.; Santiago, G.; Ferrer, M.; Guallar, V. Computationally Driven Rational Design of Substrate Promiscuity on Serine Ester Hydrolases. ACS Catal. 2021, 11, 3590–3601. [Google Scholar] [CrossRef]
  9. Jumper, J.; Evans, R.; Pritzel, A.; Green, T.; Figurnov, M.; Ronneberger, O.; Tunyasuvunakool, K.; Bates, R.; Žídek, A.; Potapenko, A.; et al. Highly accurate protein structure prediction with AlphaFold. Nature 2021, 596, 583–589. [Google Scholar] [CrossRef] [PubMed]
  10. Fischer, M. The Lipase Engineering Database: A navigation and analysis tool for protein families. Nucleic Acids Res. 2003, 31, 319–321. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Vass, M.; Shi, D.; Abualrous, E.; Chambers, J.; Chopra, N.; Higgs, C.; Kasavajhala, K.; Li, H.; Nandekar, P.; et al. Benchmarking Refined and Unrefined AlphaFold2 Structures for Hit Discovery. ChemRxiv 2022. This content is a preprint and has not been peer-reviewed. [Google Scholar] [CrossRef]
  12. Li, F.; Yuan, L.; Lu, H.; Li, G.; Chen, Y.; Engqvist, M.K.M.; Kerkhoven, E.J.; Nielsen, J. Deep learning-based kcat prediction enables improved enzyme-constrained model reconstruction. Nat. Catal. 2022, 5, 662–672. [Google Scholar] [CrossRef]
  13. Freund, G.S.; Yao, A.; Mak, W.S.; Tagkopoulos, I.; Facciotti, M.T.; Tantillo, D.J.; Siegel, J.B. Elucidating Substrate Promiscuity within the FabI Enzyme Family. ACS Chem. Biol. 2017, 12, 2465–2473. [Google Scholar] [CrossRef] [PubMed]
  14. Carbonell, P.; Faulon, J.-L. Molecular signatures-based prediction of enzyme promiscuity. Bioinformatics 2010, 26, 2012–2019. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Pertusi, D.A.; Moura, M.E.; Jeffryes, J.G.; Prabhu, S.; Walters Biggs, B.; Tyo, K.E.J. Predicting novel substrates for enzymes with minimal experimental effort with active learning. Metab. Eng. 2017, 44, 171–181. [Google Scholar] [CrossRef]
  16. Goldman, S.; Das, R.; Yang, K.K.; Coley, C.W. Machine learning modeling of family wide enzyme-substrate specificity screens. PLoS Comput. Biol. 2022, 18, e1009853. [Google Scholar] [CrossRef]
  17. Uniref. Available online: https://ftp.uniprot.org/pub/databases/uniprot/uniref/uniref50/README (accessed on 3 August 2022).
  18. Wang, J.; Yang, B.; Revote, J.; Leier, A.; Marquez-Lago, T.T.; Webb, G.; Song, J.; Chou, K.-C.; Lithgow, T. POSSUM: A bioinformatics toolkit for generating numerical sequence feature descriptors based on PSSM profiles. Bioinformatics 2017, 33, 2756–2758. [Google Scholar] [CrossRef] [Green Version]
  19. Chen, Z.; Zhao, P.; Li, F.; Leier, A.; Marquez-Lago, T.T.; Wang, Y.; Webb, G.I.; Smith, A.I.; Daly, R.J.; Chou, K.-C.; et al. iFeature: A Python package and web server for features extraction and selection from protein and peptide sequences. Bioinformatics 2018, 34, 2499–2502. [Google Scholar] [CrossRef] [Green Version]
  20. Vabalas, A.; Gowen, E.; Poliakoff, E.; Casson, A.J. Machine learning algorithm validation with a limited sample size. PLoS ONE 2019, 14, e0224365. [Google Scholar] [CrossRef]
  21. Chou, K.-C.; Shen, H.-B. MemType-2L: A Web server for predicting membrane proteins and their types by incorporating evolution information through Pse-PSSM. Biochem. Biophys. Res. Commun. 2007, 360, 339–345. [Google Scholar] [CrossRef]
  22. Zahiri, J.; Yaghoubi, O.; Mohammad-Noori, M.; Ebrahimpour, R.; Masoudi-Nejad, A. PPIevo: Protein–protein interaction prediction from PSSM based evolutionary information. Genomics 2013, 102, 237–242. [Google Scholar] [CrossRef] [PubMed]
  23. Guyon, I.; Elisseeff, A. An Introduction to Feature Extraction. In Feature Extraction; Guyon, I., Nikravesh, M., Gunn, S., Zadeh, L.A., Eds.; Studies in Fuzziness and Soft Computing; Springer: Berlin/Heidelberg, Germany, 2006; Volume 207, pp. 1–25. [Google Scholar] [CrossRef] [Green Version]
  24. Pilnenskiy, N.; Smetannikov, I. Feature Selection Algorithms as One of the Python Data Analytical Tools. Future Internet 2020, 12, 54. [Google Scholar] [CrossRef] [Green Version]
  25. Kursa, M.B.; Rudnicki, W.R. Feature Selection with the Boruta Package. J. Stat. Softw. 2010, 36, 1–13. [Google Scholar] [CrossRef] [Green Version]
  26. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature Selection: A Data Perspective. ACM Comput. Surv. 2018, 50, 1–45. [Google Scholar] [CrossRef] [Green Version]
  27. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
  28. Kubat, M. Performance Evaluation. In An Introduction to Machine Learning; Springer International Publishing: Cham, Switzerland, 2017; pp. 211–229. [Google Scholar] [CrossRef]
  29. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [Green Version]
  30. Sahigara, F.; Ballabio, D.; Todeschini, R.; Consonni, V. Defining a novel k-nearest neighbours approach to assess the applicability domain of a QSAR model for reliable predictions. J. Cheminformatics 2013, 5, 27. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Pérez-García, P.; Danso, D.; Zhang, H.; Chow, J.; Streit, W.R. Exploring the global metagenome for plastic-degrading enzymes. In Methods in Enzymology; Elsevier: Amsterdam, The Netherlands, 2021; Volume 648, pp. 137–157. [Google Scholar] [CrossRef]
  32. ModWeb. Available online: https://modbase.compbio.ucsf.edu/modweb/ (accessed on 22 July 2022).
  33. Wiederstein, M.; Sippl, M.J. ProSA-web: Interactive web service for the recognition of errors in three-dimensional structures of proteins. Nucleic Acids Res. 2007, 35, W407–W410. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Halgren, T. New Method for Fast and Accurate Binding-site Identification and Analysis. Chem. Biol. Drug Des. 2007, 69, 146–148. [Google Scholar] [CrossRef]
  35. Halgren, T.A. Identifying and Characterizing Binding Sites and Assessing Druggability. J. Chem. Inf. Model. 2009, 49, 377–389. [Google Scholar] [CrossRef]
  36. Roda, S.; Fernandez-Lopez, L.; Benedens, M.; Bollinger, A.; Thies, S.; Schumacher, J.; Coscolín, C.; Kazemi, M.; Santiago, G.; Gertzen, C.G.W.; et al. A Plurizyme with Transaminase and Hydrolase Activity Catalyzes Cascade Reactions. Angew. Chem. 2022, 134, e202207344. [Google Scholar] [CrossRef]
  37. Vidal, P.; Martínez-Martínez, M.; Fernandez-Lopez, L.; Roda, S.; Méndez-García, C.; Golyshina, O.V.; Guallar, V.; Peláez, A.I.; Ferrer, M. Metagenomic Mining for Esterases in the Microbial Community of Los Rueldos Acid Mine Drainage Formation. Front. Microbiol. 2022, 13, 868839. [Google Scholar] [CrossRef] [PubMed]
  38. Alcaide, M.; Tornés, J.; Stogios, P.J.; Xu, X.; Gertler, C.; Di Leo, R.; Bargiela, R.; Lafraya, Á.; Guazzaroni, M.-E.; López-Cortés, N.; et al. Single residues dictate the co-evolution of dual esterases: MCP hydrolases from the α/β hydrolase family. Biochem. J. 2013, 454, 157–166. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Nutschel, C.; Coscolín, C.; David, B.; Mulnaes, D.; Ferrer, M.; Jaeger, K.-E.; Gohlke, H. Promiscuous Esterases Counterintuitively Are Less Flexible than Specific Ones. J. Chem. Inf. Model. 2021, 61, 2383–2395. [Google Scholar] [CrossRef] [PubMed]
  40. Fushinobu, S.; Saku, T.; Hidaka, M.; Jun, S.-Y.; Nojiri, H.; Yamane, H.; Shoun, H.; Omori, T. Takayoshi Wakagi Crystal structures of a meta-cleavage product hydrolase from Pseudomo. Protein Sci. 2009, 11, 2184–2195. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Graphical representation of the model training process. Data was split five times into different test and training sets. The training set was then used for tuning the hyperparameters using 5-fold cross-validation (CV) while the test set was used to evaluate the trained models.
Figure 1. Graphical representation of the model training process. Data was split five times into different test and training sets. The training set was then used for tuning the hyperparameters using 5-fold cross-validation (CV) while the test set was used to evaluate the trained models.
Biomolecules 12 01529 g001
Figure 2. Matthew’s correlation coefficient (MCC) scores of the different classifiers. Ridge is the RidgeClassifier which is one of the linear models implemented in Scikit-Learn; SVM is the support vector machine; KNN is the K-nearest neighbors and EP-pred is the ensemble classifier that combined all 3 of the previous classifiers.
Figure 2. Matthew’s correlation coefficient (MCC) scores of the different classifiers. Ridge is the RidgeClassifier which is one of the linear models implemented in Scikit-Learn; SVM is the support vector machine; KNN is the K-nearest neighbors and EP-pred is the ensemble classifier that combined all 3 of the previous classifiers.
Biomolecules 12 01529 g002
Figure 3. A description of the bioprospecting workflow. A, Since there was a mix of different families in LED, first we applied an HMM profile created from the esterase dataset to clean the database and keep only esterases. B, EP-pred evaluated the remaining sequences and predicted around 500 positive hits. C, The top 100 sequences according to E-values returned by HMM in step A were isolated and analyzed according to molecular descriptors from homology modeling (HM) and Sitemap calculations. D, A final set of 10 sequences with the highest hydrophobicity and enclosure/exposure scores were gathered and sent to be validated experimentally.
Figure 3. A description of the bioprospecting workflow. A, Since there was a mix of different families in LED, first we applied an HMM profile created from the esterase dataset to clean the database and keep only esterases. B, EP-pred evaluated the remaining sequences and predicted around 500 positive hits. C, The top 100 sequences according to E-values returned by HMM in step A were isolated and analyzed according to molecular descriptors from homology modeling (HM) and Sitemap calculations. D, A final set of 10 sequences with the highest hydrophobicity and enclosure/exposure scores were gathered and sent to be validated experimentally.
Biomolecules 12 01529 g003
Figure 4. Substrate specificity of all ten hydrolases characterized in this study. The specific activity was measured for 2 meta-cleavage compounds (HOHD and HOPHD) and a set of 96 carboxylic esters (only those 89 found to be converted are shown). The proteins are organized based on their sequence similarity.
Figure 4. Substrate specificity of all ten hydrolases characterized in this study. The specific activity was measured for 2 meta-cleavage compounds (HOHD and HOPHD) and a set of 96 carboxylic esters (only those 89 found to be converted are shown). The proteins are organized based on their sequence similarity.
Biomolecules 12 01529 g004
Figure 5. The ranking of all ten characterized hydrolases in terms of promiscuity. The number of substrates converted by each of the ten hydrolases herein reported compared to that of other 145 ester hydrolases previously characterized, using the same set of esters herein tested. This figure is created from data previously reported [4] and data herein reported for the ten hydrolases.
Figure 5. The ranking of all ten characterized hydrolases in terms of promiscuity. The number of substrates converted by each of the ten hydrolases herein reported compared to that of other 145 ester hydrolases previously characterized, using the same set of esters herein tested. This figure is created from data previously reported [4] and data herein reported for the ten hydrolases.
Biomolecules 12 01529 g005
Table 1. Specific activity against p-NP esters. The results are the mean ± SD of triplicates.
Table 1. Specific activity against p-NP esters. The results are the mean ± SD of triplicates.
Specific Activity (Units mg−1)
Enzymep-NP Acetatep-NP Propionatep-NP Butyrate
AJP48854.12.96 ± 0.363.99 ± 0.251.68 ± 0.10
ART39858.12.53 ± 0.393.63 ± 0.261.48 ± 0.13
PHR82761.14.39 ± 0.442.75 ± 0.192.41 ± 0.25
WP_014900537.10.64 ± 0.022.19 ± 0.171.31 ± 0.18
WP_026140314.11.47 ± 0.053.33 ± 0.141.22 ± 0.09
WP_042877612.10.51 ± 0.012.57 ± 0.240.99 ± 0.06
WP_059541090.10.97 ± 0.022.96 ± 0.231.06 ± 0.08
WP_069226497.10.56 ± 0.012.26 ± 0.131.01 ± 0.04
WP_089515094.13.75 ± 0.174.46 ± 0.192.20 ± 0.15
WP_101198885.14.32 ± 0.145.85 ± 0.083.15 ± 0.25
Table 2. Kinetic parameters of selected hydrolases for p-NP propionate. The results are the mean ± SD of triplicates. Note: Kinetic parameters could not be determined (n.d.) because no reliable Km could be determined (low affinity for the substrate under our experimental conditions).
Table 2. Kinetic parameters of selected hydrolases for p-NP propionate. The results are the mean ± SD of triplicates. Note: Kinetic parameters could not be determined (n.d.) because no reliable Km could be determined (low affinity for the substrate under our experimental conditions).
Enzymekcat (min−1)Km (µM)
AJP48854.1n.d.n.d.
ART39858.133.1 ± 0.133.3 ± 3.1
PHR82761.12569.3 ± 6.333.7 ± 3.6
WP_014900537.146.8 ± 0.016.9 ± 1.6
WP_026140314.15.9 ± 0.014.5 ± 1.1
WP_042877612.15043.0 ± 126865.8 ± 26.1
WP_059541090.13246.8 ± 10.631 ± 3.3
WP_069226497.13775.1 ± 37.246.5 ± 11.8
WP_089515094.12060.8 ± 5.723.5 ± 3.5
WP_101198885.13452 ± 15.948.7 ± 7.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiang, R.; Fernandez-Lopez, L.; Robles-Martín, A.; Ferrer, M.; Guallar, V. EP-Pred: A Machine Learning Tool for Bioprospecting Promiscuous Ester Hydrolases. Biomolecules 2022, 12, 1529. https://doi.org/10.3390/biom12101529

AMA Style

Xiang R, Fernandez-Lopez L, Robles-Martín A, Ferrer M, Guallar V. EP-Pred: A Machine Learning Tool for Bioprospecting Promiscuous Ester Hydrolases. Biomolecules. 2022; 12(10):1529. https://doi.org/10.3390/biom12101529

Chicago/Turabian Style

Xiang, Ruite, Laura Fernandez-Lopez, Ana Robles-Martín, Manuel Ferrer, and Victor Guallar. 2022. "EP-Pred: A Machine Learning Tool for Bioprospecting Promiscuous Ester Hydrolases" Biomolecules 12, no. 10: 1529. https://doi.org/10.3390/biom12101529

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop