Next Article in Journal
Absolute Ranging with Time Delay Interferometry for Space-Borne Gravitational Wave Detection
Previous Article in Journal
Industrial Internet of Things Gateway with OPC UA Based on Sitara AM335X with ModbusE Acquisition Cycle Performance Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Collaborative Multiple Players to Address Label Sparsity in Quality Prediction of Batch Processes

1
School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi 214122, China
2
Department of Chemical and Biological Engineering, Hong Kong University of Science and Technology, Clear Water Bay, Kowloon, Hong Kong SAR 999077, China
3
School of Food Science and Technology, Jiangnan University, Wuxi 214122, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(7), 2073; https://doi.org/10.3390/s24072073
Submission received: 8 February 2024 / Revised: 14 March 2024 / Accepted: 19 March 2024 / Published: 24 March 2024
(This article belongs to the Section Chemical Sensors)

Abstract

:
For decades, soft sensors have been extensively renowned for their efficiency in real-time tracking of expensive variables for advanced process control. However, despite the diverse efforts lavished on enhancing their models, the issue of label sparsity when modeling the soft sensors has always posed challenges across various processes. In this paper, a fledgling technique, called co-training, is studied for leveraging only a small ratio of labeled data, to hone and formulate a more advantageous framework in soft sensor modeling. Dissimilar to the conventional routine where only two players are employed, we investigate the efficient number of players in batch processes, making a multiple-player learning scheme to assuage the sparsity issue. Meanwhile, a sliding window spanning across both time and batch direction is used to aggregate the samples for prediction, and account for the unique 2D correlations among the general batch process data. Altogether, the forged framework can outperform the other prevalent methods, especially when the ratio of unlabeled data is climbing up, and two case studies are showcased to demonstrate its effectiveness.

1. Introduction

Batch processes are now widely used in many high-value-added fields, ranging from food, pharmaceuticals, to the most cutting-edge silicon. The advanced control of these processes is of paramount importance for meeting overall specifications, while requiring a close oversight of certain key performance indicators (KPI) [1,2,3,4,5]. However, variables that are highly influential to them are often expensive to measure, owing to the challenges in sensor installation, analysis delays, and the overall costs [6,7,8]. To alleviate this issue, soft sensor, also known as inferential sensor, has gained popularity in recent decades, and is now recognized as an affordable and reliable prediction method that facilitates the process assessment and optimization [9,10,11,12,13].
The advantage of soft sensor lies in its avoidance of using direct hardware to measure variables, rather, using only primary measurements to enable the prediction of quality variables. Depending on the methodology, soft sensors can be divided into first-principles models and data-driven models. Because various statistical and machine learning tools have been readily available for data processing, and the data are more convenient than ever to collect, compared to the previous sophisticated model, data-driven methods are emerging as a predominant solution in the era of big data and artificial intelligence [14,15,16,17,18]. For instance, Gopakumar et al. developed a soft sensor based on deep neural network (DNN), and used it for online estimation of product quality and biomass concentration, showing more significant accuracy in streptokinase and penicillin processes than traditional SVM-based methods [19]. Albeit, with multifarious extensions to boost the model performance, the modeling of soft sensor is uniformly owed to a so-called training phase which fits the relation between the input and output data. This inevitable step generally relies on a substantial amount of data, especially the indispensable Y data, but it is expensive and scarce due to the inherent difficulty in measuring. Therefore, the deficiency of Y data results in the under-labeled issue impedes the ordinary modeling procedure of soft sensors accordingly.
To assuage the issue of label sparsity, semi-supervised learning (SSL) is deemed a feasible strategy during the modeling [20]. Semi-supervised learning uses less labeled data combined with unlabeled data to attain a better performance than using labeled data alone, which has been intensively researched in industrial applications recently [21]. Ge et al. integrated self-training into phase representative PLS models, prompting a better result than traditional PLS models given only a small portion of data was labeled from an injection molding process [22]. Jin et al. proposed an ensemble evolutionary optimization-based pseudo-labeling method (EnEOPL), wherein the Gaussian process regression (GPR) was assigned as base learner, being further enhanced by an ensemble framework to optimize and generate the pseudo-labels [23]. Esche et al. proved that when the time interval between two individual samples is noticeably too large, the SSL effect delivered by proposed deep-kernel learning is significant, as illustrated in Williams-Otto simulation and bioethanol production process [24]. However, these methods only utilized a single base learner, of which the monotone learning characteristic can obscure the real X-Y mapping that could be more perplexed underlying the process, thus easily leading to under-fitting or over-fitting issues during the modeling.
Alternatively, the idea of synthesizing multiple basic learners rather than using just one can enable an array of various perspectives when capturing the features, thereby reducing the variance across different testing datasets or conditions. Li et al. presented a semi-supervised ensemble support vector regression (SSESVR), which inferred and aggregated artificial labels into an under-labeled dataset via more than one base learner to formulate the ensemble hierarchy and facilitate the subsequent collaboration with semi-supervised learning [25]. As another viable route, co-training takes a step towards accentuating the heterogeneity by which the multiple players (base learners) can learn from the dataset [20,26]. The co-training first splits the dataset into two parts which are supposedly independent, then an individual classifier will be assigned to each and trained before screening the unlabeled samples for predicting their pseudo-labels. The pseudo-samples will be later evaluated and selected to replenish the dataset that is irrelevant to the current classifier for the next round of training, until the criteria become useless or the model converges [27]. So far, co-training has found its efficacy in many tasks such as web page classification and has become a fledgling tool in soft sensor applications. To name a few, it could be the co-training regressors (Coreg) that were brought forward by Zhou and Li, where a large amount of unlabeled data dominated the dataset and the routine was used for increasing the regression accuracy [28]. In addition, Bao et al. applied PLS instead of kNN as the basic model, showing better results in simulation and real plant data [29]. Besides, Tang et al. presented a co-training style kernel extreme learning machine (ELM), which deployed two ELMs with designated kernel tricks to label samples in each dataset, demonstrating good capability in handling under-labeled and industrial datasets [30]. However, previous methods based on multiple learners mostly stick to the structure itself, like co-training, without seeking changes in the structure or taking more into account regarding the characteristics of the batch process itself.
The aforementioned work laid a solid foundation in terms of soft sensor modeling and insufficient label treatment. However, there all still several limitations that have not been focused on before: (1) The first are the unique data characteristics brought by the scenario itself. A batch process, wherein the soft sensor is being applied, typically encompasses multiple phases and transitions within and across the batches. This mandates that, not only are some nonlinearities supposed to exist in the process data, but some patterns, known as 2D dynamics, that evolve with time and batch direction would also hamper the data analysis and the subsequent soft sensor modeling, hence vitiating the model capability and its accuracy. (2) Though co-training and the renewed Coreg proved to outperform in sparse label learning (regression), their essences are still based on the original two-learner co-training framework. Regardless of the ensemble nature, if the in-situ base learner is inappropriate or insufficient to capture the heterogeneous features, then a singleton type of base learners may deliver an inferior cooperation when later predicting pseudo-labels, hence degrading the practical significance in multi-phase batch processes. (3) Traditional regression models often either use deep learning for prediction, or directly utilize co-training-based models to predict the final variables. However, in batch process scenarios that lack labeled data, deep learning is limited by the inability to obtain sufficient data labels, and models based on co-training are difficult to capture the complex characteristics of the batch process. Therefore, we believe that only a framework that integrates co-training and deep learning can better solve the problem of quality variable prediction in the current scenario. Therefore, in order to devise a procedure which is geared towards using sparse labels when modeling soft sensors for batch processes, an instructional multi-player framework is showcased in this research to mitigate the scarcity of quality variables and to account for the nonlinearities and dynamics that hinder the modeling and application. The main contributions are thus highlighted as follows:
  • Rather than using only two base learners, we investigate the efficient number of different base learners used in co-training, which paves the way for capturing multi-channeled input features that are leveraged in pseudo-labels generation;
  • Once the pseudo-labels complete the data augmentation, a sliding window is skillfully embedded preceding the feature engineering, to account for the unique 2D dynamics of batch processes;
  • Leveraging the pseudo-labels inferred by the local feature similarity, a deep learning interface named 2D-GSTAE, is further connected to synthesize all the perspectives presented by the previous base learners, promoting a more comprehensive relationship between the input process data and the online estimated output.
Therefore, a framework in the name of co-training with multiple players in deep soft sensor modeling (CO-MP-DSSM) is presented in this work, by which other fields may also benefit given appropriate deep synthesizers and modifications. The framework first adds pseudo-labels to the data through collaborative multi-player learning, then simultaneously extracts two-dimensional dynamic features in the time direction and batch direction through a two-dimensional dynamic sliding window, and finally utilizes deep learning to fuse the features and predict quality variables. The remaining sections of the paper are ordered as follows. In Section 2, a review of the co-training and Coreg is given. In Section 3, the details of the CO-MP-DSSM framework are dissected. In Section 4, case studies on simulation and real fermentation processes are provided to evaluate the prediction performance. At the end, the summary and outlook of soft sensor in batch processes are given in the last section.

2. Preliminary

2.1. Original Co-Training

The concept of co-training is originally proposed by Blum and Mitchel [27]. Its implementation is shown in Figure 1, where L is labeled dataset, L 1 and L 2 are the different views of training set, h 1 and h 2 are the different classifier, U donates the unlabeled dataset and U represents the random selected unlabeled dataset, t is the test dataset and y represents the prediction result. Co-training firstly divides the labeled dataset into two supposedly independent and redundant views, L 1 and L 2 , then trains different classifiers h 1 and h 2 on them and utilizes the prediction results to expand each other’s training set. The reduced part of U will be supplemented from U . After k times of the above iterative training, h 1 and h 2 are finally used to classify the test set. Co-training has proven useful in classification scenarios with few labeled samples, but due to the aforementioned limits, it can be still inferior when directly used in soft sensor modeling.

2.2. Co-Training Regressors (Coreg)

As an effective semi-supervised learning paradigm, co-training has received a lot of attention since its novel concept, but related research is only based on classification problems. Therefore, Zhou and Li proposed a co-training style regression model, co-training regressors (Coreg) [28]. Since it is very similar to co-training, the structure of the Coreg algorithm is also shown in Figure 1 and the difference is marked as red, where r 1 and r 2 are changed to the different regressors. Both r 1 and r 2 are based on kNN regressor but use different Minkowski distances. The Minkowski distance is calculated as follows,
M i n k o w s k y p ( x a , x b ) = ( i = 1 d x a i x b i p ) 1 p
where x a and x b are two samples in the multi-dimensional space, d represents the total dimension of the sample and p denote the distance order. Minkowski distance is a metric only when p is not less than 1. Generally speaking, when the value p is large the distance metric is more sensitive to samples, and when the value p is small, the distance metric is more robust.
During the training phase, Coreg algorithm first duplicates the labeled dataset L into two copies, L 1 and L 2 , respectively, as training sets for the two regressors r 1 and r 2 to initialize them. Coreg then works in a similar way to co-training, but the specific steps are different. When it iterates, each time the two regressors first traverse all the data in U , predict the label of unlabeled data U j , and add it to the training set before retraining the regressors. Then the regressor is used to predict the k nearest neighbors to U j from the labeled dataset, and the loss is calculated as follows:
U j = Argmax U j U ( x i Ω ( y i r a ( x i ) ) 2 / Ω x i Ω ( y i r a ( x i ) ) 2 / Ω )
where r a and r a are one of the regressors and the retrained regressor after adding the predicted label, and Ω is the k nearest neighbors to U j in the labeled dataset. The U j with maximum positive loss and its predicted label will be added to another labeled dataset, while no data will be added if without positive loss. If neither L 1 or L 2 change, the iteration will end early. Finally, r 1 and r 2 will be retrained using the changed L 1 and L 2 . After training is completed, Coreg predicts the test set t as follows:
y = ( r 1 ( x ) + r 2 ( x ) ) / 2

3. Proposed Method

In this paper, in order to fully utilize the small amount of labeled data as well as the large amount of unlabeled data, and predict the quality variables of the batch processes, we propose a co-training-based semi-supervised deep learning model CO-MP-DSSM. The structure of CO-MP-DSSM is shown in Figure 2, with the details presented explicitly as follows. Suppose the process data and quality data of batch processes have been collected and organized as X I × J × K and Y I × 1 × L , where I indicates the number of batches, J represents the number of variables, K is the number of sampling numbers for process variables, and L denotes the sampling number for quality variable. In CO-MP-DSSM, the data are processed and modeled as follows:

3.1. Collaborative Multiple-Player Structure to Infer the Labels in Individual Perspectives

Based on previous Coreg, we proposed a collaborative scheme with multiple players to mitigate the scarcity of the labeled data. The pseudocode for this algorithm is shown as Algorithm 1. Regarding this, we firstly divide the batch processes data into labeled data and unlabeled data and copy the labeled data into several copies, corresponding to the predefined number of regressors.
A l g o r i t h m   1   Collaborative   Multiple   Players   Learning
I n p u t :   l a b e l e d   d a t a s e t   L ,   u n l a b e l e d   d a t a s e t   U ,
number   of   the   nearest   neighbors   k ,
the   maximum   number   of   iteratios   N ,
distance   order   p 1 , p 2 , , p n
P r o c e d u r e :
L 1 L ;   L 2 L ;   ;   L n L ;   number   of   iterations   n u m 0
P 1 k N N ( L 1 , k , p 1 ) ;   P 2 k N N ( L 2 , k , p 2 ) ; P n k N N ( L n , k , p n )   #   Initialize   all   the   datasets   and   regressors
w h i l e   n u m < N   d o
n u m n u m + 1
f o r   j { 1 , 2 , , n } d o #   Perform   iteration   over   each   regressor
f o r   each   x U   d o       #   Perform   iteration   over   each   unlabeled   data
y ^ r j ( x )         #   Predict   the   pseudo-label   for   x
P j k N N ( { ( x , y ^ ) } L j , k , p j )     #   Train   new   regressor   using   data   containing   x   and   y ^
a G e t N e i g h b o r s ( x , k , L j )   #   Get   the   neighbors   of   x   from   labeled   data
Δ x i x i a ( ( y i P j ( x i ) 2 ( y i P j ( x i ) 2 ) / | a |     #   Calculate   the   difference   before   and   after   adding
e n d   f o r
i f   e x i s t   Δ x i < 0   t h e n
x i Argmin x i U Δ x i ; y ^ i P j ( x i )       #   Select   the   optimal   x   and   y ^   from   unlabeled   data
π j { ( x i , y ^ i ) } , U U π j   #   Remove   x   from   the   unlabeled   data
e l s e   π
e n d   i f
e n d   f o r
π s π 1 π 2 π n
L 1 L 1 ( π s π 1 ) ; L 2 L 2 ( π s π 2 ) ; ; L n L n ( π s π n )   #   Add   the   optimal   x   and   y ^   to   each   other   labeled   data
i f   n e i t h e r   o f   L 1 , L 2 , , L n   c h a n g e   t h e n   e x i t
e l s e
P 1 k N N ( L 1 , k , p 1 ) ; P 2 k N N ( L 2 , k , p 2 ) ;
P n k N N ( L n , k , p n ) #   Retrain   the   regressors   with   updated   data
e n d   i f
e n d   w h i l e
e n d   P r o c e d u r e
O u t p u t : r ( x ) ( r 1 ( x ) + r 2 ( x ) + + r n ( x ) ) / n   is   used   to   predict   the   pseudo   label   of   test   set   t
After initializing the regressors using labeled data, we iteratively train the regressors. For each iteration, we train each regressor in turns, using them to traverse all data in the unlabeled dataset. For each unlabeled datum, we first predict its pseudo-label through the previous regressor P , and retrain it using the dataset after adding the data and its pseudo-label to obtain the regressor P . In order to evaluate the performance of the regressor after adding the data and its pseudo-labels, we approximately take the k nearest neighbors of the data in the labeled data and compare the RMSE predicted by P and P . The loss of the regressor changes as follows:
x ^ = Argmin x ^ U ( x i a ( y i P ( x i ) ) 2 x i a ( y i P ( x i ) ) 2 / a )
where a denotes the k nearest neighbors of the data x ^ . According to previous research, we believe that the smaller the loss, the better the effect of adding this data to improve the prediction performance of the regressor. For this training, we obtain the unlabeled data corresponding to the smallest negative loss and its pseudo-label and add it to the training set of other regressors. If no loss is negative, we believe that no data can improve the performance of the regressor and we do not add any data to other training sets. Iterative training is repeated until the number of times reaches the upper limit or all regressors are unable to select unlabeled data that can improve the prediction effect. After training is complete, we use unlabeled training data as input and predict their labels. The label of each datum is calculated as:
y = ( P 1 ( x ) + P 2 ( x ) + + P n ( x ) ) / n
where P 1 , P 2 , …, P n represent the regressors trained using different Minkowski distances. Finally, the training set is reshaped so that it has the same shape as at the beginning.

3.2. Preprocess and 2D Slide Window Preceding the Modeling

To reduce the impact on the model caused by differences in data values and units, for train x, train y, and test x we standardize the three-dimensional data. For instance, the data with shape I × J × K is first transformed into J × I × K . Next, we divide the data into a matrix of dimension J × I × K and standardize it using z-score by row, respectively. For each matrix, the standard deviation is:
σ = i = 1 N ( x i x ¯ ) 2 / N
where N is the number of elements in the matrix, and x ¯ represents the mean of the elements. In the matrix each element x is normalized as:
s = ( x x ¯ ) / σ
where s denotes the result of standardizing. Afterwards, the matrix is transformed back into I × J × K .
In order to extract two-dimensional dynamic features from time and batch directions simultaneously, we utilize 2D sliding windows to process the data. For train x, train y and test x, we use a 2D sliding window to first slide in the time direction moment by moment until the end, then slide to the next position in the batch direction and slide in the time direction from the beginning. The details of the 2D sliding window method are illustrated in Figure 3, where the data undergo dimensionality reduction after being processed through the sliding window. For example, if the data of I × J × K is slid using the sliding window of p × q , we obtain data with shape of p × q × J × I p + 1 × K q + 1 , which is a five-dimensional tensor. The size of the sliding window is selected according to the data dimension and can be adjusted according to the experimental results. Since tensors with too high dimensions cannot be trained using deep neural networks, and too many dimensions are meaningless, we first spliced the data in the sliding window batch by batch, and then arranged the sliding windows in sliding order to obtain processed data ( p × q ) × J ×   [ I p + 1 × K q + 1 ] .

3.3. Deep Learning: Fusing and Prediction

According to our previous research, GSTAE achieves better results for batch process quality variable prediction compared to traditional SAE for quality prediction in batch processes [31]. Compared to traditional SAE, GSTAE not only utilizes gate unit to integrate the feature of data in different layers, but also adds quality-related information in the pre-training stage to allow the network to extract more features related to quality variables [32]. The pseudocode of GSTAE is shown in Appendix A. So, in the end, the processed data is used to train GSTAE model, whose training phase consists of supervised pre-training stage and fine-tuning stage. In the pre-training stage, GSTAE first attempts to extract more abstract features by mapping the input data to a hidden layer with fewer neurons, and then reconstructs the input through the hidden layer to reduce the loss of information. GSTAE reconstructs the data as:
[ x ˜ , y ˜ ] T = f ˜ ( W 2 f ( W 1 x + b 1 ) + b 2 )
where x is input matrix, x ˜ and y ˜ are the reconstruction results of x , and y , W 1 and W 2 are the weight matrix, b 1 and b 2 are the bias vector. The network is trained layer-by-layer in the above order. The number of neurons in the first layer of the network is equal to the number of input variables, and then the number of neurons in each layer gradually decreases. The prediction of quality variables in each layer can avoid the loss of information related to quality variables while obtaining more abstract process variable characteristics. In the fine-tuning stage, the model predicts the label and tries to fit the real label, so the loss is calculated as
L o s s = i = 1 N y i y ˜ i 2
where N is the total amount of data. After training is completed, the network inputs sample features and weights the predicted labels of each layer through the gate unit. The gate unit controls the impact of this layer of information on the final result through the weight matrix, bias vector and activation function. Then the predictions of each layer are added in sum to obtain the final predicted label. The label is calculated as
y ˜ = i = 1 n σ ( W g i × h i + b g i ) tanh ( W o i × h i + b o i )
where n represents the layer of GSTAE, W g and W o are the weight matrix of the gate, b g and b o are the bias vector of the gate.

3.4. Evaluation Indicators

In order to facilitate the comparison of prediction effects between models, we choose the most widely used performance evaluation indices RMSE and R2 as indicators of model prediction effects. The RMSE and R2 are calculated as
R M S E = i = 1 n ( y i y ˜ i ) 2 / n
R 2 = i = 1 n ( y ˜ i y ¯ ) 2 / i = 1 n ( y i y ¯ ) 2
In general, a decrease in RMSE and an increase in R2 indicate an improved predictive performance of the model.

4. Case Study

4.1. Penicillin Fermentation Simulation Case

Penicillin is one of the most commonly used antibiotics. Most of it is synthesized through fermentation and purification. The fermentation process of penicillin is a typical fed-batch process, which has strong nonlinearity and two-dimensional dynamic characteristics. To prove the effectiveness of the model, we utilize penicillin fermentation simulation platform PenSim V2.0 to generate simulated data [33]. Since the data generated by PenSim V2.0 are close to the real situation, and most research in related fields also uses this platform, we also use this platform for simulation experiments.

4.1.1. Experiment Design in Simulation Case

By using the simulation platform, we generated a total of 100 batches of data, of which 75 batches were used as training sets and 25 batches were used as test sets. The total fermentation time is set to 400 h, and the sampling interval is set to 1 h. In order to simulate the two-dimensional dynamic features in real process, we set initial conditions as follows. For batch direction, pH is set to gradually grow from 4.95 to 5.2, temperature is set to gradually grow from 297.5 to 298.5. Besides, substrate concentration is set to 15, agitator power is set to 30, carbon dioxide concentration is set to 0.5, aeration rate is set to 8.6 and culture volume is set to 100. The above variables of different batches fluctuate within about 1% of the set value. The remaining variables are set to the default values of Pensim V2.0. The variables selected for modeling are displayed in Table 1. Because the data generated by the Pensim is too ideal, we add a certain amount of noise to the data to make it closer to the real result.

4.1.2. Parameter Settings in Simulation Case

To showcase the superiority and evaluate the performance of models, SAE, SIAE, GSTAE, 2D-GSTAE, SS-GSTAE, SS-2D-GSTAE, Coreg, and CO-MP-DSSM are constructed. Among them, 2D-GSTAE is the CO-MP-DSSM without a collaborative multiple-player structure, Coreg is the collaborative multiple-player structure without the other part, and SS-2D-GSTAE is 2D-GSTAE combined with traditional semi-supervised learning structurally. The parameters of the model and comparison models are set as follows. The structure of CO-MP-DSSM is set as {11, 9, 7, 4}, where 11 is the number of neurons in the first layer must be consistent with the number of variables, and 9, 7, 4 are the numbers of neurons in hidden layers. According to the test, the structure above has given the best prediction performance among all structures. The learning rate of the model is set as 0.01. The trade-off coefficient λ is set to 0.5. For comparison models, the structure of SAE, SIAE, GSTAE, SS-GSTAE, 2D-GSTAE are also set as {11, 9, 7, 4}. Such identical parameter settings not only make all models perform well, but also make comparisons between models relatively fair. The learning rate of SAE, SIAE, GSTAE, SS-GSTAE, 2D-GSTAE, SS-2D-GSTAE is set as 0.01. The trade-off coefficient λ of GSTAE, SS-GSTAE, 2D-GSTAE, SS-2D-GSTAE is set to 0.5. The number of regressors in co-training is set to 2 for Coreg and 3 for CO-MP-DSSM. The sizes of the two-dimensional sliding windows of 2D-GSTAE, SS-2D-GSTAE, and CO-MP-DSSM in batch and time are set to 7 and 5. The number of nearest neighbors of kNN in Coreg and CO-MP-DSSM is set to 15.

4.1.3. Results and Analysis in Simulation Case

The results of models are presented in Table 2, where the proposed model and best performance in each ratio is highlighted in bold. The summary of comparative results among different methods is displayed in Figure 4 with standout methods represented in different colors. All networks were initialized with ten different random seeds, and their result indicators were averaged to obtain the final experimental results. The sampling ratio in Table 2 denotes the proportion of labeled data in total data. Since this case is a simulation experiment, the effect prediction of the model with small noise and other interference factors is relatively stable, thereby facilitating a clearer comparison of the results influenced by the model’s structure. It is not difficult to find from the charts and tables that traditional methods such as SAE and SIAE are not fitted well for batch processes, while the model based on GSTAE achieves significantly better results. Since SAE and SIAE only consider x-related information in the pre-training stage, the models’ ability to extract quality-related information is limited. Moreover, the simple structure further limits the model’s nonlinear feature extraction capabilities, resulting in poor prediction performance. Therefore, it is evident that the structure of the basic model determines the prediction effect of the model to a certain extent.
The remaining models achieve similar results when ample labeled data is available, but the gap of effects gradually appear as the labeled data are reduced. When labeled data are scarce, the effects of GSTAE and 2D-GSTAE, experience the most pronounced deterioration in performance. This is because the model that is structured as supervised learning cannot effectively utilize unlabeled data, leading to significant performance degradation as the quantity of labeled data diminishes. Compared with the former, SS-GSTAE and SS-2D-GSTAE utilize semi-supervised learning to deal with unlabeled data and achieve certain results in dealing with data with labeled data accounting for more than 1/25. In order to highlight the effect of SS-2D-GSTAE, we mark its results with a blue line in Figure 4. However, since labeled data may be scarcer in actual situations, the models’ effect of semi-supervised learning will also decline significantly once it exceeds a specific threshold. In addition, no matter at which ratio, the model with a 2D sliding window consistently outperform those without, due to their enhanced ability to capture the dynamic characteristics of batch processes.
However, when labeled data are scarce, the situation is different. Compared with other semi-supervised learning methods, the models based on co-training emphasize the heterogeneity of the structural design. So when faced with a large amount of labeled data the effect is close to traditional semi-supervised learning based method, and it can avoid underfitting and overfitting to achieve better performance when the labels are scarce. Among them, because CO-MP-DSSM integrates multi-player co-training, 2D dynamic feature extraction and deep learning model learning, it can not only obtain more trustworthy pseudo-labels, but also avoid potential issues caused by the too simple structure of the base model by using deep learning. Therefore, CO-MP-DSSM achieves the smallest RMSE and the biggest R2 among all the models. In order to highlight the effect of CO-MP-DSSM, we mark its results with a red line in Figure 4.

4.2. Real Lactic Acid Bacteria Fermentation Case

In order to further verify the effectiveness of the model, we conducted real lactic acid bacteria fermentation experiments. Lactobacillus plantarum is a member of the genus Lactobacillus and has the potential to be widely used in the food and fermentation industries. It has antioxidant, antibacterial and antiviral potential, helps maintain the balance of intestinal flora, and promotes the health of the digestive system. To evaluate the effectiveness of the model, we selected Lactobacillus plantarum HuNHHMY71L1 as the fermentation strain.

4.2.1. Experiment Design in Real Case

In this real case, we obtained a total of 20 batches of data through fermentation, of which 15 batches were used as training sets and 5 batches were used as test sets. According to previous reference, the total fermentation time was set to 8 h, the sampling interval of process variables was set to 1 min, while due to conditional restrictions the sampling interval of quality variable was set to 30 min. Supplementary culture medium was added to the fermenter after 2 h of fermentation. In addition, MRS was used as the culture medium, the fermentation temperature was set to 37 °C allowing a margin of error of 0.2, the pH was set to 6 allowing a margin of error of 0.02, the culture volume was set to 3 L, and the stirring speed is was to 100 rmp. Due to changes in external conditions during the fermentation process and operational errors during the experiment, the fermentation process had two-dimensional dynamic characteristics like other fed-batch processes. Figure 5 depicts the instruments utilized in our fermentation process, along with corresponding schematic diagrams of their specific structures. The measured variables are presented in Table 3.

4.2.2. Parameter Settings in Real Case

To compare the performances of the models, the same models as in the previous case were constructed, and some hyperparameters were adjusted according to the data. The parameters of the model and comparison models were set as follows. The structure of models was set as {7, 6, 5, 3}, where 7 is the number of neurons in the first layer that must be consistent with the number of variables, and 6, 5, 3 are the number of neurons in hidden layers. The number of regressors in co-training was set to 2 for Coreg and 3 for CO-MP-DSSM. The sizes of the two-dimensional sliding windows of 2D-GSTAE, SS-2D-GSTAE and CO-MP-DSSM in batch and time were set to 3 and 3. The number of nearest neighbors of kNN in Coreg and CO-MP-DSSM is set to 10.

4.2.3. Result and Analysis in Real Case

The results of models are shown in Table 4, where the proposed model and best performance in each ratio is highlighted in bold. Figure 6 employs boxplots to compare the predictive performance of different models and the stability of their predictions. The sampling ratio in Table 4 denotes the proportion of labeled data in total data. Due to the limitations of sampling in the actual process, only approximately 1/30 of the total dataset could be labeled. Therefore, we can only compare the performance of various models when labels are scarce. We have specifically chosen to analyze results for three different sampling ratios as further reduction in labeled data would render the experiment less meaningful. Moreover, variations in the intervals of labeled data could compromise the experimental design. Additionally, it is important to note that the effect of the model may exhibit relative instability due to the significant influence of external factors in the actual operational process.
Despite the limited amount of labeled data, the quality-related information cannot be ignored for the final prediction result, resulting in the suboptimal performance of SAE and SIAE. GSTAE has demonstrated stable performance owing to its structural advantages, while models SS-GSTAE and 2D-GSTAE have achieved better results through the incorporation of semi-supervised and 2D sliding window structures, respectively. However, unlike the previous case, SS-2D-GSTAE did not achieve better results than SS-GSTAE and 2D-GSTAE. This may be due to the smaller number of batches limiting the size selection of the sliding window, and the smaller number of labeled data resulting in a lack of reliable label information in some windows. Therefore, when dealing with actual cases, effective combination of model structures is more important than simple stacking structures. Since only a simple base learner is used and the strong nonlinearity and 2D dynamic characteristics of the batch processes are not taken into account, sole reliance on Coreg does not yield satisfactory outcomes. Through the effective combination of multi-player co-training, 2D dynamic sliding windows and deep learning models, CO-MP-DSSM leverages their strengths while mitigating weaknesses, thereby effectively learning the characteristics of the batch process and achieving the smallest RMSE and biggest R2 among all the models. In order to highlight the effect of CO-MP-DSSM, we mark its results with an orange boxplot in Figure 6.
Figure 7 and Table 5 show another reason behind the superior performance of CO-MP-DSSM over Coreg, and provide another perspective on why we chose to use three regressors in this case. When we used CO-MP-DSSM with 1–6 regressors for prediction tasks, we observed that the model utilizing three regressors not only outperformed other configurations but also exhibited greater stability in prediction outcomes across most scenarios. Furthermore, in most cases when the number of regressors is less than the best number, increasing the number of regressors will improve the model effect, while when the number of regressors is greater than the best number, decreasing the number of regressors will improve the model performance. Compared with primitive 2D-GSTAE, CO-MP-DSSM using three regressors is not only stable, but also performs better in most cases, especially when the number of labeled data is relatively trustworthy. In addition to the accuracy brought by multiple players, we believe that even using less accurate pseudo-labels can also improve the prediction effect of the model, so CO-MP-DSSM achieves far better results than 2D-GSTAE. So, in this article we used CO-MP-DSSM with three players for training.

5. Conclusions

Label scarcity is an actual problem faced by many deep-learning-based methods when working on chemical soft sensor modeling. In this paper, a collaborative multiple-player combined framework, named CO-MP-DSSM, has been put forward to address the label sparsity in the quality prediction of batch processes. First, dissimilar to the conventional scheme of co-training where there used to be only two players, the number of effective players has been studied to generate more significant pseudo labels, even when the ratio of true labels is very scarce. Afterwards, a 2D sliding window is used on the batch data to capture the correlations in both the time and batch directions. Finally, for illustration, the processed data is put into the GSTAE model for the up-scale extraction of features and quality prediction. Experimental results on two datasets demonstrate that CO-MP-DSSM achieves better results than other models without this technique. However, since each of the multi-players is supposed to use kNN to iterate over all the unlabeled data, the computation is relatively expensive in the training phase compared with other methods. In the future, researchers may explore more efficient perspectives for players to measure the similarity between the data, aiming to refine methods that can achieve better results while minimizing computational resources.

Author Contributions

Conceptualization, L.Z., Z.Z. and J.Z.; methodology, L.Z., Z.Z. and J.Z.; software, L.Z.; validation, L.Z. and Z.Z; formal analysis, L.Z. and Z.Z.; investigation, L.Z. and Z.Z; resources, L.Z. and J.Z.; data curation, L.Z.; writing—original draft preparation, L.Z. and Z.Z.; writing—review and editing, L.Z. and Z.Z.; visualization, L.Z. and Z.Z.; supervision, Z.Z., J.Z., H.W. and Z.X.; project administration, J.Z.; funding acquisition, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Jiangsu Province (Grant No. BK20210452), the National Natural Science Foundation of China (Grant No. 62103168) and the Fundamental Research Funds for the Central Universities (JUSRP622034).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The pseudo code of GSTAE, used as an example of DSSM in this research, is appended as Algorithm A1. For 2D mechanism and its solution, it may refer readers to our earlier conference [31].
A l g o r i t h m   A 1   Gated   Stacked   Target- Related   Autoencoder
I n p u t : training   feature   x ,   training   label   y
   AElist   l i s t ,   pre-training   epochs   p e ,   fine-tuning   epochs   f e ,
pre-training   batch   size   p s ,   fine-tuning   batch   size   f s
trade-off   c o e f f i c i e n t   λ
P r o c e d u r e :
f o r   j = 0   to   l e n ( l i s t ) d o #   Perform   iteration   over   each   layer   in   pre-training   stage
f o r   i =   0   to   p e   d o     #   Conduct   pre-training
l o s s f m s e L o s s
f o r   e a c h   p s ( x ) x , p s ( y ) y   d o     #   Perform   iteration   over   each   batch
X ^ , y ^ f ˜ ( W i 2 f ( W i 1 x + b i 1 ) + b i 2 )   #   Reconstruction   and   prediction
l o s s l o s s f ( x , x ^ ) + l o s s f ( y , y ^ ) × λ     #   Calculate   the   loss
t r a i n   l i s t ( i ) l o s s
e n d   f o r
e n d   w h i l e
e n d   w h i l e
f o r   i =   0   to   f e   d o #   Perform   iteration   over   each   batch   in   fine-tuning   stage
l o s s f m s e L o s s
f o r   e a c h   f s ( x ) x , f s ( y ) y   d o       #   Perform   iteration   over   each   batch
y ^ 0
f o r   e a c h   l i s t ( i ) l i s t   d o   #   Integrating   information   from   various   layers
y ^ y ^ + G a t e ( f ( W i 1 x + b i 1 ) )
e n d   f o r
l o s s l o s s f ( y , y ^ )
f o r   e a c h   l i s t ( i ) l i s t   d o #   Fine-tune   parameters   in   each   layer
f i n e - t u n i n g   l i s t ( i ) l o s s
e n d   f o r
e n d   f o r
e n d   w h i l e
e n d   P r o c e d u r e
O u t p u t : GSTAE   network

References

  1. Zhang, Z.; Zhu, J.; Zhang, S.; Gao, F. Process monitoring using recurrent Kalman variational auto-encoder for general complex dynamic processes. Eng. Appl. Artif. Intell. 2023, 123, 106424. [Google Scholar] [CrossRef]
  2. Zhu, J.; Gao, F. Improved nonlinear quality estimation for multiphase batch processes based on relevance vector machine with neighborhood component variable selection. Ind. Eng. Chem. Res. 2018, 57, 666–676. [Google Scholar] [CrossRef]
  3. Curreri, F.; Patanè, L.; Xibilia, M.G. Soft sensor transferability: A survey. Appl. Sci. 2021, 11, 7710. [Google Scholar] [CrossRef]
  4. Sun, Q.; Ge, Z. A survey on deep learning for data-driven soft sensors. IEEE Trans. Ind. Inform. 2021, 17, 5853–5866. [Google Scholar] [CrossRef]
  5. Zheng, J.; Shen, F.; Ye, L. Improved mahalanobis distance based JITL-LSTM soft sensor for multiphase batch processes. IEEE Access 2021, 9, 72172–72182. [Google Scholar] [CrossRef]
  6. Zhang, Z.; Jiang, Q.; Wang, G.; Pan, C.; Cao, Z.; Yan, X.; Zhuang, Y. Neural networks-based hybrid beneficial variable selection and modeling for soft sensing. Control Eng. Pract. 2023, 139, 105613. [Google Scholar] [CrossRef]
  7. Shen, F.; Zheng, J.; Ye, L.; Ma, X. LSTM soft sensor development of batch processes with multivariate trajectory-based ensemble just-in-time learning. IEEE Access 2020, 8, 73855–73864. [Google Scholar]
  8. Hua, L.; Zhang, C.; Sun, W.; Li, Y.; Xiong, J.; Nazir, M.S. An evolutionary deep learning soft sensor model based on random forest feature selection technique for penicillin fermentation process. ISA Trans. 2023, 136, 139–151. [Google Scholar] [CrossRef]
  9. Zhu, X.; Rehman, K.U.; Wang, B.; Shahzad, M. Modern soft-sensing modeling methods for fermentation processes. Sensors 2020, 20, 1771. [Google Scholar] [CrossRef]
  10. Brunner, V.; Siegl, M.; Geier, D.; Becker, T. Challenges in the development of soft sensors for bioprocesses: A critical review. Front. Bioeng. Biotechnol. 2021, 9, 722202. [Google Scholar] [CrossRef]
  11. Qiu, K.; Wang, J.; Wang, R.; Guo, Y.; Zhao, L. Soft sensor development based on kernel dynamic time warping and a relevant vector machine for unequal-length batch processes. Expert Syst. Appl. 2021, 182, 115223. [Google Scholar] [CrossRef]
  12. Wang, J.; Qiu, K.; Wang, R.; Zhou, X.; Guo, Y. Development of soft sensor based on sequential kernel fuzzy partitioning and just-in-time relevance vector machine for multiphase batch processes. IEEE Trans. Instrum. Meas. 2021, 70, 2509110. [Google Scholar] [CrossRef]
  13. Ji, C.; Ma, F.; Wang, J.; Sun, W. Profitability related industrial-scale batch processes monitoring via deep learning based soft sensor development. Comput. Chem. Eng. 2023, 170, 108125. [Google Scholar] [CrossRef]
  14. Yuan, X.; Wang, Y.; Yang, C.; Gui, W. Stacked isomorphic autoencoder based soft analyzer and its application to sulfur recovery unit. Inf. Sci. 2020, 534, 72–84. [Google Scholar] [CrossRef]
  15. Yuan, X.; Feng, L.; Wang, K.; Wang, Y.; Ye, L. Deep learning for data modeling of multirate quality variables in industrial processes. IEEE Trans. Instrum. Meas. 2021, 70, 2509611. [Google Scholar] [CrossRef]
  16. Jiang, Q.; Wang, Z.; Yan, S.; Cao, Z. Data-driven soft sensing for batch processes using neural network-based deep quality-relevant representation learning. IEEE Trans. Artif. Intell. 2022, 4, 602–611. [Google Scholar] [CrossRef]
  17. Ren, J.-C.; Liu, D.; Wan, Y. VMD-SEAE-TL-Based Data-Driven soft sensor modeling for a complex industrial batch processes. Measurement 2022, 198, 111439. [Google Scholar] [CrossRef]
  18. Sun, Y.-N.; Qin, W.; Xu, H.-W.; Tan, R.-Z.; Zhang, Z.-L.; Shi, W.-T. A multiphase information fusion strategy for data-driven quality prediction of industrial batch processes. Inf. Sci. 2022, 608, 81–95. [Google Scholar] [CrossRef]
  19. Gopakumar, V.; Tiwari, S.; Rahman, I. A deep learning based data driven soft sensor for bioprocesses. Biochem. Eng. J. 2018, 136, 28–39. [Google Scholar] [CrossRef]
  20. Ge, Z. Semi-supervised data modeling and analytics in the process industry: Current research status and challenges. IFAC J. Syst. Control 2021, 16, 100150. [Google Scholar] [CrossRef]
  21. Van Engelen, J.E.; Hoos, H.H. A survey on semi-supervised learning. Mach. Learn. 2020, 109, 373–440. [Google Scholar] [CrossRef]
  22. Ge, Z.; Song, Z.; Gao, F. Self-training statistical quality prediction of batch processes with limited quality data. Ind. Eng. Chem. Res. 2013, 52, 979–984. [Google Scholar] [CrossRef]
  23. Jin, H.; Li, Z.; Chen, X.; Qian, B.; Yang, B.; Yang, J. Evolutionary optimization based pseudo labeling for semi-supervised soft sensor development of industrial processes. Chem. Eng. Sci. 2021, 237, 116560. [Google Scholar] [CrossRef]
  24. Esche, E.; Talis, T.; Weigert, J.; Rihm, G.B.; You, B.; Hoffmann, C.; Repke, J.-U. Semi-supervised learning for data-driven soft-sensing of biological and chemical processes. Chem. Eng. Sci. 2022, 251, 117459. [Google Scholar] [CrossRef]
  25. Li, Z.; Jin, H.; Dong, S.; Qian, B.; Yang, B.; Chen, X. Semi-supervised ensemble support vector regression based soft sensor for key quality variable estimation of nonlinear industrial processes with limited labeled data. Chem. Eng. Res. Des. 2022, 179, 510–526. [Google Scholar] [CrossRef]
  26. Ning, X.; Wang, X.; Xu, S.; Cai, W.; Zhang, L.; Yu, L.; Li, W. A review of research on co-training. Concurr. Comput. Pract. Exp. 2023, 35, e6276. [Google Scholar] [CrossRef]
  27. Blum, A.; Mitchell, T. Combining labeled and unlabeled data with co-training. In Proceedings of the Eleventh Annual Conference on Computational Learning Theory, Madison, WI, USA, 24–26 July 1998; pp. 92–100. [Google Scholar]
  28. Zhou, Z.-H.; Li, M. Semi-supervised regression with co-training. In Proceedings of the IJCAI, Nineteenth International Joint Conference on Artificial Intelligence, Edinburgh, UK, 30 July–5 August 2005; pp. 908–913. [Google Scholar]
  29. Bao, L.; Yuan, X.; Ge, Z. Co-training partial least squares model for semi-supervised soft sensor development. Chemom. Intell. Lab. Syst. 2015, 147, 75–85. [Google Scholar] [CrossRef]
  30. Tang, Q.; Li, D.; Xi, Y. Soft sensor modeling based on cotraining-style kernel extreme learning machine. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; pp. 4028–4033. [Google Scholar]
  31. Zhao, L.; Zhu, J.; Zhang, Z.; Xie, Z.; Gao, F. A Novel Semi-supervised Two-dimensional Dynamic Soft Sensor for Quality Prediction in Batch Processes. In Proceedings of the 2023 5th International Conference on Industrial Artificial Intelligence (IAI), Shenyang, China, 21–24 August 2023; pp. 1–6. [Google Scholar]
  32. Sun, Q.; Ge, Z. Gated stacked target-related autoencoder: A novel deep feature extraction and layerwise ensemble method for industrial soft sensor application. IEEE Trans. Cybern. 2020, 52, 3457–3468. [Google Scholar] [CrossRef]
  33. Birol, G.; Ündey, C.; Cinar, A. A modular simulation package for fed-batch fermentation: Penicillin production. Comput. Chem. Eng. 2002, 26, 1553–1565. [Google Scholar] [CrossRef]
Figure 1. Dissected diagram of co-training and Coreg (red) algorithm. (The major difference lies in that co-training has its dataset split by two views from the raw data, while Coreg simply duplicates the data into two copies. Besides, the base learners of them are termed classifiers and regressors, respectively).
Figure 1. Dissected diagram of co-training and Coreg (red) algorithm. (The major difference lies in that co-training has its dataset split by two views from the raw data, while Coreg simply duplicates the data into two copies. Besides, the base learners of them are termed classifiers and regressors, respectively).
Sensors 24 02073 g001
Figure 2. Procedure of proposed CO-MP-DSSM framework: (1) details explained in Section 3; (2) & (3) Section 3.1; (4) Section 3.2; (5) Section 3.3.
Figure 2. Procedure of proposed CO-MP-DSSM framework: (1) details explained in Section 3; (2) & (3) Section 3.1; (4) Section 3.2; (5) Section 3.3.
Sensors 24 02073 g002
Figure 3. Flowchart of the employed 2D slide window (As depicted above, the window first slides forward along the time direction, then shifts into the beginning of the next batch once the current batch ends.)
Figure 3. Flowchart of the employed 2D slide window (As depicted above, the window first slides forward along the time direction, then shifts into the beginning of the next batch once the current batch ends.)
Sensors 24 02073 g003
Figure 4. Aggregated results of average RMSE and R2 of models compared in the case of Pensim dataset.
Figure 4. Aggregated results of average RMSE and R2 of models compared in the case of Pensim dataset.
Sensors 24 02073 g004
Figure 5. Instruments for lactic acid bacteria fermentation (the left side is the actual picture, and the right side is the schematic diagram. The equipment on the picture corresponds to each other).
Figure 5. Instruments for lactic acid bacteria fermentation (the left side is the actual picture, and the right side is the schematic diagram. The equipment on the picture corresponds to each other).
Sensors 24 02073 g005
Figure 6. Comparative performances (boxplot) of RMSE in real lactic acid bacteria fermentation dataset.
Figure 6. Comparative performances (boxplot) of RMSE in real lactic acid bacteria fermentation dataset.
Sensors 24 02073 g006
Figure 7. Comparative performances (boxplot) of RMSE from models with various players and label ratios, based on the real lactic acid bacteria fermentation dataset. (When number of players is set to 1, 2D-GSTAE is structurally equal to CO-MP-DSSM, except for slight higher RMSE due to no pseudo labels involved in its training. For better illustration, we repeat the plot for 2D-GSTAE across different situations, to highlight the accuracy and stability improvements by the proposed method.)
Figure 7. Comparative performances (boxplot) of RMSE from models with various players and label ratios, based on the real lactic acid bacteria fermentation dataset. (When number of players is set to 1, 2D-GSTAE is structurally equal to CO-MP-DSSM, except for slight higher RMSE due to no pseudo labels involved in its training. For better illustration, we repeat the plot for 2D-GSTAE across different situations, to highlight the accuracy and stability improvements by the proposed method.)
Sensors 24 02073 g007
Table 1. Selected variables of penicillin fermentation process.
Table 1. Selected variables of penicillin fermentation process.
NumberVariableUnit
1aeration rateL/h
2agitator powerW
3substrate feed rateL/h
4substrate feed temperatureK
5substrate concentrationg/L
6culture volumeL
7carbon dioxide concentrationg/L
8pH-
9temperatureK
10generated heatkcal
11cold water flow rateL/h
ypenicillin concentrationg/L
Table 2. Tabulated average RMSE and R2 of models compared in penicillin fermentation process. (The model in bold is the method proposed in this article, and the number in bold is the best one in the column of data)
Table 2. Tabulated average RMSE and R2 of models compared in penicillin fermentation process. (The model in bold is the method proposed in this article, and the number in bold is the best one in the column of data)
Method5:110:120:125:140:150:1
RMSER2RMSER2RMSER2RMSER2RMSER2RMSER2
SAE0.06280.98140.07690.97370.10240.95410.12300.93190.26370.69800.7977−2.5510
SIAE0.05210.98780.05550.98620.09020.96570.10290.95280.27910.62030.6989−1.5870
GSTAE0.01830.99860.02070.99820.03330.99530.04230.99230.10600.95070.21190.8036
SS-GSTAE0.01570.99900.01840.99860.02930.99640.03630.99450.10230.95490.15680.8964
2D-GSTAE0.01190.99940.01560.99900.02930.99640.03450.99500.09270.96220.20270.8202
SS-2D-GSTAE0.01100.99950.01150.99940.02300.99770.02980.99510.06450.98190.11470.9402
Coreg0.01970.99840.02470.99750.03440.99510.03860.99380.05400.98780.07320.9777
CO-MP-DSSM0.01880.99850.02280.99780.03090.99600.03660.99430.05180.98880.07180.9786
Table 3. Selected variables of lactic acid bacteria fermentation process.
Table 3. Selected variables of lactic acid bacteria fermentation process.
NumberVariableUnit
1temperatureK
2pH-
3dissolved oxygen-
4agitator rater/min
5acid supplementmL
6base supplementmL
7substrate supplementmL
ylactic acid bacteria concentration-
Table 4. Tabulated average RMSE and R2 of models compared in lactic acid bacteria fermen-tation process. (The model in bold is the method proposed in this article, and the number in bold is the best one in the column of data)
Table 4. Tabulated average RMSE and R2 of models compared in lactic acid bacteria fermen-tation process. (The model in bold is the method proposed in this article, and the number in bold is the best one in the column of data)
Method30:160:1120:1
RMSER2RMSER2RMSER2
SAE0.72040.91860.78040.90310.97590.8574
SIAE0.82150.89510.94290.86570.95970.8513
GSTAE0.77040.91110.83650.89650.92420.8746
SS-GSTAE0.68320.93140.75150.91630.87340.8880
2D-GSTAE0.69620.92870.76220.91430.98430.8576
SS-2D-GSTAE0.92660.87271.00610.84961.06040.8341
Coreg0.97310.86110.95580.86600.97220.8614
CO-MP-DSSM0.55930.95380.56980.95240.74550.9184
Table 5. The average prediction RMSE and R2 of co-training methods with different number of players in lactic acid bacteria fermentation process. (The model in bold is the method proposed in this article, and the number in bold is the best one in the column of data)
Table 5. The average prediction RMSE and R2 of co-training methods with different number of players in lactic acid bacteria fermentation process. (The model in bold is the method proposed in this article, and the number in bold is the best one in the column of data)
Number of Players30:160:1120:1
RMSER2RMSER2RMSER2
10.60320.94650.66300.93490.74270.9188
20.58630.94900.60000.94710.75400.9165
30.55930.95380.56980.95240.74550.9184
40.57920.95050.62610.94220.73010.9218
50.58970.94870.71770.92390.78070.9106
60.63590.94060.63290.94120.76490.9142
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, L.; Zhang, Z.; Zhu, J.; Wang, H.; Xie, Z. Collaborative Multiple Players to Address Label Sparsity in Quality Prediction of Batch Processes. Sensors 2024, 24, 2073. https://doi.org/10.3390/s24072073

AMA Style

Zhao L, Zhang Z, Zhu J, Wang H, Xie Z. Collaborative Multiple Players to Address Label Sparsity in Quality Prediction of Batch Processes. Sensors. 2024; 24(7):2073. https://doi.org/10.3390/s24072073

Chicago/Turabian Style

Zhao, Ling, Zheng Zhang, Jinlin Zhu, Hongchao Wang, and Zhenping Xie. 2024. "Collaborative Multiple Players to Address Label Sparsity in Quality Prediction of Batch Processes" Sensors 24, no. 7: 2073. https://doi.org/10.3390/s24072073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop