Next Article in Journal
Remote Power Generation for Applications to Natural Gas Grid: A Comprehensive Market Review of Techno-Energetic, Economic and Environmental Performance
Next Article in Special Issue
Quantitative Evaluation of Imbibition Damage Characteristics of Foaming Agent Solutions in Shale Reservoir
Previous Article in Journal
Analysis of Light Utility Vehicle Readiness in Military Transportation Systems Using Markov and Semi-Markov Processes
Previous Article in Special Issue
Steam Cavity Expansion Model for Steam Flooding in Deep Heavy Oil Reservoirs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Oil Saturation during Water and Gas Injection Using Controllable Convolutional Long Short-Term Memory

1
College of Computer Science and Technology, China University of Petroleum (East China), Qingdao 266580, China
2
Research Institute of Exploration and Development, Tarim Oilfield Company, PetroChina, Korla 841000, China
*
Author to whom correspondence should be addressed.
Energies 2022, 15(14), 5063; https://doi.org/10.3390/en15145063
Submission received: 16 June 2022 / Revised: 3 July 2022 / Accepted: 9 July 2022 / Published: 11 July 2022
(This article belongs to the Special Issue Enhanced Oil Recovery for Unconventional Oil and Gas Reservoirs)

Abstract

:
Oil saturation is a kind of spatiotemporal sequence that changes dynamically with time, and it is affected not only by the reservoir properties, but also by the injection–production parameters. When predicting oil saturation during water and gas injection, the influence of time, space and injection–production parameters should be considered. Aiming at this issue, a prediction method based on a controllable convolutional long short-term memory network (Ctrl-CLSTM) is proposed in this paper. The Ctrl-CLSTM is an unsupervised learning model whose input is the previous spatiotemporal sequence together with the controllable factors of corresponding moments, and the output is the sequence to be predicted. In this way, future oil saturation can be generated from the historical context. Concretely, the convolution operation is embedded into each unit to describe the interaction between temporal features and spatial structures of oil saturation, thus the Ctrl-CLSTM realizes the unified modeling of the spatiotemporal features of oil saturation. In addition, a novel control gate structure is introduced in each Ctrl-CLSTM unit to take the injection–production parameters as controllable influencing factors and establish the nonlinear relationship between oil saturation and injection–production parameters according to the coordinates of each well location. Therefore, different oil saturation prediction results can be obtained by changing the injection–production parameters. Finally, experiments on real oilfields show that the Ctrl-CLSTM comprehensively considers the influence of artificial controllable factors such as injection–production parameters, accomplishes accurate prediction of oil saturation with a structure similarity of more than 98% and is more time efficient than reservoir numerical simulation.

1. Introduction

The oil recovery efficiency of an oilfield depends on the geological environment of reservoir and the current level of oil extraction technology. Typically, the oil recovery efficiency is only 30% to 50% in most areas of China, and the remaining 50% to 70% of the oil remains in reservoirs [1]. Therefore, appropriate measures must be implemented to enhance the oil recovery efficiency, including the water and gas injection, which are widely implemented as the secondary oil recovery technology. To improve oil recovery, oilfield experts continuously simulate different production processes by changing injection–production parameters such as water or gas injection rate and water or gas injection timing [2]; sometimes, they also change the well pattern structure, such as adding a water injection well or shutting down a gas injection well. Oil saturation provides a degree of assistance to oilfield experts in their decision-making, which directly represents the oil content in different positions of the reservoir. Clarifying the oil saturation means finding the direction of oilfield development strategy adjustment [3], which has important guiding significance for timely adjustment of the injection–production parameters in the oilfield production site.
Since the mid-20th century, researchers from all over the world have paid extensive attention to the issues related to oil saturation prediction [4]. After continuous research studies, various methods for predicting oil saturation have been proposed, such as seismic analysis, reservoir engineering, well testing and reservoir numerical simulation [5,6,7,8]; these traditional methods have made a significant contribution to enhanced reservoir recovery. However, in order to predict oil saturation more quickly and accurately, up-to-date methods should be introduced in the oilfield.
In recent years, with the rapid development of artificial intelligence and big data, the oil and gas resources informatization development has been continuously carried out in the field of petroleum engineering [9]. Those previous conventional studies and the large quantity of accumulated data provide strong supports for deep learning methods to predict oil saturation. Gu et al. used the fully connected long short-term memory [8] (FC-LSTM) network to predict oil saturation, and the influence of different injection–production working systems on oil saturation was considered in the training stage. FC-LSTM handles the temporality of the data naturally, but cannot model the spatial structure of the reservoir sufficiently. However, both temporality and spatiality should be taken into account when building neural networks [10]. Zhang et al. used a convolutional neural-network-based model to predict oil saturation maps [11], which considered various reservoir properties including permeability and active grid. They transformed each of these reservoir properties into a two-dimensional array and made these arrays into a stacked multilayer array, then they used this stacked multilayer array to generate oil saturation maps. Nevertheless, convolutional neural networks could not learn the temporal features effectively, so the whole training process did not model the temporal features of oilfield production processes.
In fact, the treatment of spatiotemporal features is a general problem, which has already been researched by existing studies. ConvLSTM [12], a convolution long short-term memory network, embeds a convolution operation into the LSTM network and replaces the full connection layer with convolution; it obtained good results in many kinds of spatiotemporal sequence prediction tasks [13]. In ConvLSTM, the memory state is only transmitted in the temporal direction, so the spatial modeling capability can still be further improved [14,15]. Based on the prediction method of the spatiotemporal long short-term memory network ST-LSTM [16], a longitudinal spatial memory state transition path was introduced into the network, which effectively reduced the gradient disappearance between neurons in longitudinal layers, so that the final prediction results could widely consider the hierarchical spatiotemporal features.
Oil saturation data consist of special spatiotemporal sequence data, which need to consider simultaneously the effect of geological properties and injection–production parameters [17]. On this basis, this paper proposes a prediction method based on a Ctrl-CLSTM, which adopts a dual-memory state transition mechanism to achieve the unified modeling of the temporal and spatial characteristics of oil saturation. Moreover, we primarily consider the effect of artificial and operable conditions on the oil saturation. By adding a new control gate structure to the network, the model learns more about the influence of the injection–production parameters on oil saturation, which improves the prediction accuracy and makes the prediction process more in line with physical laws. Compared with the traditional numerical simulation, the Ctrl-CLSTM adjusts the injection–production parameters more quickly and maintains an equivalent prediction accuracy. Analyzing the oil saturation generated from different combinations of injection–production parameters, a theoretical reference can be provided to oilfield experts when designing injection–production working systems.

2. Spatiotemporal Sequence Prediction

A spatiotemporal sequence describes the multidimensional grid information, which is applied to various spatiotemporal sequence prediction tasks, such as short-term precipitation prediction, traffic flow prediction, marine pollution trajectory prediction, etc. In addition to changing with time, a spatiotemporal sequence is also interdependent between adjacent spatial grids [18]. The oil saturation is also a multidimensional grid sequence that changes dynamically with time, and the distribution states of adjacent spatial locations affect each other, so it can be defined as a spatiotemporal sequence, and its temporal and spatial characteristics must be learned simultaneously.
Our purpose is to capture the complex dynamic physical laws from the spatiotemporal system of oil saturation continuously observed, and approximately estimate the future trend of oil saturation. A generalized approach for predicting oil saturation using deep learning can be summarized below: the first is to create a training sample set through the reservoir numerical simulation software tool and obtain a large number of grid data of size m × n with spatial correlation. For a multilayer reservoir, each training sample contains r oil saturation maps; then, the oil saturation at time t can be recorded as X t R m × n × r ( r = 1 , 2 , 3 , ) . Let X α : β be the oil saturation sequence of time period α β , in Equation (1), the sequence X α : β is the input of the neural network, and the sequence X ^ β : γ is the most approximate estimation of the future oil saturation sequence X β : γ .
X ^ β : γ = a r g m a x X β : γ R m × n × r p ( X β : γ | X α : β )
Of note, the inherent characteristics of a reservoir, such as permeability distribution, formation thickness, boundary dip angle and other factors, have a great impact on the distribution of oil saturation. Moreover, in order to improve oil recovery, oilfield experts usually design a variety of well locations, as well as injection–production working systems, including different parameters such as displacement agents, injection amounts and injection timing. Changing the value of the injection–production parameters will also affect the state of the crude oil in the reservoir; for example, increasing the bottom pressure will make the crude oil have a stronger flow capacity in the reservoir pores [19]. In summary, oil saturation data are special spatiotemporal sequence data, which are affected by both reservoir environment and artificial controllable injection–production working systems in the development process. It is generally believed that the physical property parameters of a reservoir cannot be controlled by humans; in view of this, the change of oil saturation caused by injection–production working systems was considered mainly. The composition of the injection–production working system P is shown in Equation (2). Although the effect of reservoir parameters was not considered much, the Ctrl-CLSTM is an unsupervised learning model and its input contains historical oil saturation, from which reliable potential reservoir properties information can be learned.
P = [ T 1 , , T n ; G 1 , , G n ; W 1 , , W x ; Q 1 , , Q y ]
where n , x and y represent the number of gas injection wells, water injection wells and production wells, respectively, T i and G i represent the gas injection timing and gas injection speed, respectively, W i indicates the water injection rate and Q i indicates the bottom hole pressure of the production wells ( i = 1 , 2 , 3 , ) .
In order to ensure the value of each injection–production parameter evenly covered the entire value range, the Latin hypercube sampling [20] was used to generate the experimental data. Some examples of injection–production parameters of the Tarim oilfield in China are shown in Table 1, where each row represents an injection–production working system that lasts for ten years, and the results files are generated by numerical simulation software semiannually.
For instance, in the first row, T 1 = 3.5 means that the gas injection well G 1 is opened in the 3.5th year, and its gas injection rate is 93,459 m 3 /d, W 1 = 124 indicates that the water injection volume of water injection well W 1 is 124 m 3 /d, and Q 1 = 54 indicates that the bottom hole flow pressure of production well Q 1 is set to 54 MPa.

3. Controllable Convolutional Long Short-Term Memory Network

The Ctrl-CLSTM constructed in this paper takes the historical spatiotemporal sequence and the corresponding controllable factors as the inputs and predicts the future spatiotemporal sequence. We innovatively introduced controllable factors in the spatiotemporal sequence prediction task to specifically deal with a special sequence such as oil saturation. The core part of the Ctrl-CLSTM is the dual-memory state transmit mechanism, which realizes the unified modeling of the spatiotemporal features. In addition, the control gate structure is embedded in all nodes of the neural network, which enhances the modeling ability of the network for controllable factors such as well location distribution, injection–production parameters, etc., controlling the change of the spatiotemporal sequence effectively.

3.1. Architecture of Ctrl-CLSTM

For spatiotemporal sequence prediction tasks, different types of features are encoded into different networks architectures. LSTM is a variant of the recurrent neural network, which is suitable for sequence data prediction with strong temporal dynamics. Its transmission from input state to the hidden state uses the matrix multiplication of the fully connected layer, and the feature state is transferred in the time dimension. A convolutional neural network performs better in dealing with grid data with spatial correlation [21]. ConvLSTM [12] replaces the full connection layer in LSTM units with a convolution layer, which solves the problem that traditional LSTM cannot describe spatial structure to a certain extent. However, the memory state of ConvLSTM is only transmitted along the temporal direction, and the neurons between the vertical layers are independent of each other. The spatial features of ConvLSTM become abstract with the depth of the convolution operation, and the ability to maintain the detailed features is not enough. On this basis, ST-LSTM [16] adds a spatial memory state transmitted between longitudinal layers, which effectively reduces the gradient disappearance of spatial features and alleviates its abstraction. Finally, ST-LSTM deeply fuses the temporal memory state and spatial memory state to learn hierarchical spatiotemporal features widely. The Ctrl-CLSTM adopts this dual-memory state-transfer mechanism like ST-LSTM, and innovatively embeds the control gate structure in each neural unit, making the controllable factors as an additional input.
The network structure of the Ctrl-CLSTM is shown in Figure 1, where l represents the number of layers, t represents the time, the bottom layer is the input unit and the top layer is the output unit. The horizontal transfer path encodes the temporal state C t l , with arrows pointing to the next moment, and the vertical transfer path encodes the spatial state M t l , with arrows pointing to a deeper level of feature abstraction. The spatial state of the vertical transfer path is transmitted to the bottom unit of the next moment after reaching the top unit, which effectively reduces the gradient vanishing phenomenon. Apart from that, we can learn the memory sates C t l and M t l from memory networks, in which they can be seen as additional computational steps [22].

3.2. Units of Ctrl-CLSTM

The role of controllable factors is to intervene in the evolution of a spatiotemporal sequence and affect its development direction; different controllable factors contribute to different degrees of prediction results. As mentioned in Section 1, oil saturation is affected by many factors, especially in reservoirs with strong heterogeneity. Due to uneven distribution of oil layers and large differences in reservoir permeability, the influence of the injection–production parameters is particularly prominent [23,24]. Therefore, we select the injection–production parameters as the primary controllable factor. By introducing the control gate structure, the Ctrl-CLSTM learns the dynamic mechanism of injection–production parameters to a certain extent, integrates injection–production parameters and oil saturation and extracts their characteristics to capture the influence of the injection–production parameters on oil saturation changes.
As shown in Figure 2, the state transition path of a Ctrl-CLSTM unit is divided into two parts, including the temporal memory state C t l (Blue) and the spatial memory state M t l (Orange). In the stage of data preprocessing, the injection–production parameters should be combined with the well location map, and the geographical spatial position of the working wells is crucial to the production. Therefore, the control gate structure k t (yellow) is embedded in the spatial memory state transition path. This control gate structure increases the number of nonlinear neurons involved in the state transition calculation in each unit node, which is equivalent to expanding the memory state transition path in the spatial dimension, and conducive to improving the correlation of memory states at different spatial layers. At the same time, it enhances the modeling ability of the model for injection–production parameters and well location map, so that the network can learn the control effect of the injection–production parameters on oil saturation. To date, there are four gate structures on the M t l path, that is, the input gate i t , which selectively retains the input passed in from the previous neuron, focusing on recording the important information; the renew gate g t , which determines the content to be written; the forget gate f t , which selectively forgets the input passed in from the previous neuron, forgetting the unimportant information; and the novel control gate k t . In this way, it can be distinguished from the control signal of the temporal memory state C t l . The specific state transition method of the Ctrl-CLSTM is as follows.
Temporal memory state transition path:
g t = t a n h ( W g ( S t P t ) + W h g H t 1 l ) i t = σ ( W i ( S t P t ) + W h i H t 1 l ) f t = σ ( W f ( S t P t ) + W h f H t 1 l ) C t l = f t C t 1 l + i t g t
Spatial memory state transition path:
g t = t a n h ( W g ( S t P t ) + W m g M t l 1 ) i t = σ ( W i ( S t P t ) + W m i M t l 1 ) f t = σ ( W f ( S t P t ) + W m f M t l 1 ) k t = t a n h ( W k ( S t P t ) + W m k M t l 1 ) M t l = f t M t l 1 + i t g t + i t k t
The output gate o t further seamlessly fuses the control signals from both directions:
o t = σ ( W o ( S t P t ) + W h o H t 1 l + W c o C t l + W m o M t l )
Equations (3)–(5) are the state transition path of the Ctrl-CLSTM, where S t is the oil saturation at time t, P t is the injection–production parameters corresponding to time t and ⊕ links S t and P t in parallel. They establish a matrix with the same size as the oil saturation, then mark the water injection wells, gas injection wells and production wells on the matrix according to the well position coordinates; afterwards, they record the production parameters of the corresponding wells, and finally get a dual-channel matrix in parallel with the oil saturation. ( S t P t ) is the input of the Ctrl-CLSTM, including the oil saturation, well location information and injection–production parameters. ∗ means the convolution operation, ⊙ is the Hadamard product [25], which is the multiplication of the corresponding elements of two matrices of the same order. σ represents a sigmoid activation function, t a n h is a nonlinear hyperbolic activation function, both of which are applied between neurons in two adjacent layers, making the neural network have nonlinear mapping capability. W ( . ) is the connection weight corresponding to each Ctrl-CLSTM gate.
H t l = o t t a n h ( W ( 1 × 1 ) [ C t l , M t l ] )
Hidden state H t l is the output of a Ctrl-CLSTM unit at time t and depends on the nonlinear combination of C t l and M t l . Equation (6) reduces the dimension of the hidden state through a 1 × 1 convolution layer and a t a n h activation function, resulting in H t l having the same dimension as C t l and M t l . In particular, the bottom unit receives the memory state of the top unit at the previous time: M t l = 0 = M t 1 l = 3 .
As shown in Figure 1, the temporal memory state C t l is transmitted horizontally, the spatial memory state M t l is transmitted longitudinally and the whole is transmitted in a zigzag manner. In order to make the temporal memory state and the spatial memory state update their nonlinear functions independently at any time, the two should remain orthogonal during training. Therefore, in Equation (7), a decoupling loss function L d e c o u p l e is defined through the cosine similarity function, so that the two transmit their own characteristic parameters in their respective latent spaces independently without interfering with each other, where W d e c o u p l e represents the convolution kernel of 1 × 1 , , c is the matrix point multiplication, and . c represents the L2 regularization operation after flattening the features, which is calculated according to the channel c.
Δ C t l = W d e c o u p l e ( i t g t ) Δ M t l = W d e c o u p l e ( i t g t + i t k t ) L d e c o u p l e = t l c | Δ C t l , Δ M t l c | Δ C t l c · Δ M t l c

4. Experiments and Results

In this paper, an actual reservoir officially developed in the Tarim oilfield since 1999 was selected as the research object. The data such as water content, gas-oil ratio and static pressure have a good history matching. The reservoir model consists of 471 × 421 × 7 grids, and each grid is 10 m × 10 m × 10 m in size. There are 21 production wells and 16 injection wells in this reservoir, 12 of which are water injection wells and the other 4 are gas injection wells. Figure 3 shows the locations of the wells on the permeability distribution map, where 4 triangles mark the locations of gas injection wells, 12 pentagrams mark the locations of water injection wells, and 21 dots indicate the locations of production wells. We simulated 15 well pattern structures, each with 300 injection–production working systems, for a total of 4500 injection–production working systems. As for the value of injection–production parameters, we used the Latin hypercube sampling method to sample uniformly within a certain range. The bottom hole pressure of a production well was at least 28 MPa, the injection volume of a water injection well was in the range of 0–150 m 3 /d and the injection volume of a gas injection well was in the range of 0–120,000 m 3 /d. Finally, all injection–production working systems were fitted by the numerical simulation technology as samples. The number of samples was sufficient, which was suitable for neural network training and learning.

4.1. Constructing the Dataset

Making the data suitable for our model required the following steps:
(1)
Latin hypercube sampling (LHS): Sampling is often present in experiments, and a good sampling strategy can make our model’s sample point distribution more reasonable. We adopted LHS to generate uniformly distributed injection–production parameters. LHS is a stratified random sampling that allows efficient sampling from the distribution interval of the variables, ensuring a comprehensive coverage of the value range of each injection–production parameter. A LHS X n in [0, 1) is defined by the following equation:
x n , j = π j ( n 1 ) + U n , j N
Among Equation (8), x n , j is the jth component of sample point X n , π j is a permutation of [0, N − 1], uniformly chosen at random, N is the total number of X n and U n , j is uniformly distributed in [0, 1).
(2)
We adjusted the oil saturation data generated by the numerical simulation to a matrix of size 471 × 421. Then, Equation (9) normalized the injection–production parameters to reduce the dimensional gap between different well data and help to reduce the loss function value, where x is the parameter to be normalized, x m a x and x m i n are the maximum and minimum parameters of this working system and x is the result after normalization.
x = x x m i n x m a x x m i n
(3)
In order to utilize the well location information comprehensively, we marked and distinguish the gas injection wells, water injection wells and production wells on the well location map according to the well location coordinates, corresponding to the positions of the triangles, pentagrams and dots in Figure 3. The injection–production parameters were written at the corresponding coordinates, then we linked well location map and oil saturation in parallel to form a dual-channel matrix. In order to balance the training duration and the prediction accuracy, we discarded the area around the oilfield where the oil content was always zero, and only the effective values in the matrix were retained during the training process. Finally, the size of the dual-channel matrix was adjusted to 256 × 256 × 2, which reduced the occupation rate of the training display RAM and sped up the convergence.
(4)
According to the method proposed by Wang [15,16], we regarded the 10 parallel dual-channel matrices as an entire spatiotemporal sequence, including five historical observations (input of the model) and five future prediction results (output of the model), to avoid data interference between different working systems.
(5)
Dataset division: we selected 200 groups of injection–production working systems from every well pattern structure as the training set, 50 groups as the validation set, and 50 groups as the test set, which evenly covered all types of well pattern structures.

4.2. Implementation Details

We observed that the number of layers of the Ctrl-CLSTM had a great influence on the final performance. After repeated trials, we finally set three stacked layers in the Ctrl-CLSTM, with 128 channels for every memory state, to maintain a balance between the training efficiency and prediction quality. Unless otherwise specified, we used the ADAM optimizer [26] to train the model, which is one of the most common optimization algorithms and has been validated in a large number of deep neural networks. The initial learning rate was set to 0.001 and the convolution kernel size was set to 5 × 5. Equation (10) is the loss function used, which is the sum of the reconstruction loss function and the decoupling loss function; the deviation of the generated image from the real image is defined by the reconstruction loss function, where X t ^ and X t are the predicted oil saturation and the actual image, respectively, in each time step t, . 2 2 is the square operation and λ is a constant.
L o s s = t = β t = γ X t ^ X t 2 2 + λ L d e c o u p l e
We also noted that the selection of batch size had a great impact on the convergence speed of the loss function; therefore, we selected a minibatch of four, that is, each training iteration trained four sequences randomly at once, and the training was stopped after 50,000 training iterations. All intelligent models were set with similar hyperparameters and the experimental code was written in PyTorch, which ran on a GeForce RTX™ 3090 graphics card.

4.3. Evaluation Metrics

We selected the evaluation metrics widely adopted in the tasks of spatiotemporal sequence prediction to evaluate the trained models: the structural similarity index measure (SSIM) [27], evaluating the similarity between the predicted results and the real pictures (the value range is (0, 1) and higher is better).
S S I M = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
Among Equation (11), μ x and μ y are the mean intensity of the real picture x and the predicted result y, σ x y is the covariance of x and y, σ x and σ y are their standard deviations and C 1 and C 2 are the constant to avoid a zero denominator.
The second metric was the mean squared error (MSE), estimating the absolute pixel level error of two images (lower is better); in Equation (12), m and n are the dimensions of the images and x i j and y i j are the pixel values of the real image and the predicted image, respectively.
M S E = 1 m n i = 1 m j = 1 n ( x i j y i j ) 2
The third metric was the peak signal-to-noise ratio (PSNR) [28], evaluating the pixel-wise similarity between the predicted image and the real image (higher is better). Generally speaking, a PSNR over 40 dB means that the picture has a low level of distortion.
P S N R = 10 l o g 10 ( 255 2 / M S E ( x , y ) )
PSNR is one of the most widely used objective evaluation index of images, but it does not consider the visual recognition and perception characteristics of the human eyes, and the evaluation results are often different from the subjective perception of human beings. So we introduced the learned perceptual image patch similarity (LPIPS) [29], which aligns better with human perceptions (lower is better).
L P I P S = l 1 H l W l h , w w l ( x h w l y h w l ) 2 2
In Equation (14), Zhang et al. [29] extracted feature stacks x h w l and y h w l from L layers and unit-normalized them in the channel dimension. The vector w l was used to scale the activations channel-wise and H , W are the size of the feature stack.

4.4. Comparative Analysis of Oil Saturation Prediction Results of Different Networks

In this section, we compare and analyze the oil saturation predicted by different networks. The input of each model consisted of the previous oil saturation sequence (the sequence in Figure 4a), the injection–production parameters and the well location distribution of a corresponding period. Figure 4b shows the numerical simulation results, which is regarded as the ground truth X t ^ for comparison with other advanced spatiotemporal sequence prediction models, including the traditional long short-term memory network FC-LSTM [8], the TrajGRU [14], which adopts an encoding–predicting network structure able to learn subnetwork parameters and improve the overall network state, and the ST-LSTM [16], which introduces spatial memory flow. Some experimental results of each model on the test set are presented in Figure 5. Each row of Figure 5a shows the predicted oil saturation sequence X t of the Ctrl-CLSTM, ST-LSTM, TrajGRU and FC-LSTM, for a total of five oil saturation distributions at 360, 720, 1080, 1440 and 1800 days. Figure 5b is also from day 360 to day 1800, and it shows the error ( X t ^ X t ) between the numerical simulation result and the prediction of each model.
It is evident in Figure 5 that the oil saturation predicted by the Ctrl-CLSTM has the highest clarity, and the absolute value of the error is less than 0.04. As the time step of the forecast grows, the performance of other models becomes progressively worse, especially in the boundary where the oil content varies significantly. All of them indicate that, by embedding the control gate structure in each unit, the Ctrl-CLSTM has better learning capability and simulates the changing process of oil saturation realistically.
The average quantitative results on each evaluation metric of different models were calculated and are displayed in Table 2, from which we can draw some conclusions: the Ctrl-CLSTM has the best performance in each metric, with the SSIM exceeding 98% and the MSE decreasing from 36.61 to 21.37. LPIPS is closer to human perceptual judgment, and the Ctrl-CLSTM has minimal LPIPS indicating it has excellent power to generate high-fidelity images. With the proposed control gate structure, there is a more accurate grasp of the effect of injection–production parameters on oil saturation. Figure 6 provides the comparison results of corresponding times; as the prediction time step increases, although all models tend to deteriorate, the Ctrl-CLSTM deteriorates more slowly, and the final prediction remains highly accurate and outperforms other models.
When the Ctrl-CLSTM is applied to predict oil saturation sequence, an appropriate number of model parameters should be chosen. To obtain satisfactory experimental results, the neural network was set to two, three, and four layers separately, and the training results are shown in Table 3. It can be seen that the model performs best when three layers of Ctrl-CLSTM units are designed. When the model uses two stacked layers, the prediction effect is the worst, because only two layers of neurons can not remember the long-term variation of oil saturation. When the model uses a four-layer stack structure, the effect is also reduced, because there are too many neural network parameters at this time, which is prone to overfitting, resulting in a better effect on the training set, but a worse effect on the test set.

4.5. Results on Different Injection–Production Parameters

In the case of oil–gas–water three-phase seepage, it is more complicated to formulate a reasonable injection–production working system. Through the control gate, the dynamic physical laws between injection–production parameters and oil saturation can be captured. To verify this point of view, we took the same oil saturation as input sequence, only adjusting the injection–production parameters, then we predicted and observed the variations of oil saturation in the future. The sequence in Figure 7a and the injection–production working system of the corresponding period were the input of the Ctrl-CLSTM, Figure 7b shows the predicted results when the original injection–production working system was applied, and Figure 7c shows the predicted results when a different injection–production working system was applied. The first row is the numerical simulation result ( X t ^ ), the second row is the Ctrl-CLSTM prediction result ( X t ), the third row is the absolute value of the error ( X t ^ X t ) between the two images and the prediction time is 360, 720, 1080, 1440, 1800 and 2160 days from left to right according to columns. When using the original injection–production working system, the maximum error is 0.14, and the error in most areas is less than 0.07. After changing the injection–production working system, the maximum error is 0.2, and the error in most areas is less than 0.1. It can be seen that with the control gate, the final state of oil saturation is changed with a high prediction accuracy when changing the injection–production parameters. Therefore, the Ctrl-CLSTM can obtain different prediction results through the control gate when changing the injection–production parameters, and the injection–production parameters can be adjusted reasonably based on this reference, which can improve the production potential of oil wells.

5. Performance on Another Reservoir Dataset

To demonstrate the universality of the Ctrl-CLSTM, we performed experiments on another reservoir dataset generated by a numerical simulation. In this reservoir model, there were 3000 oil saturation sequences, and each sequence contained 10 images of size 50 × 50. We took the first four images as input sequence and the last six images as prediction result. In addition, these sequences had their respective injection–production parameters. In Figure 8, (a) is the input sequence of (c), (b) is the input sequence of (d); similarly, the injection–production working system in the corresponding period was also one of the inputs to the network. The rows represent the numerical simulation result ( X t ^ ), the prediction result ( X t ) of the Ctrl-CLSTM, and the error ( X t ^ X t ), while the columns indicate the number of days to predict oil saturation from 360 days to 2160 days.
After consecutive experiments, it can be seen that the errors between both sequences are less than 0.1 in most regions, indicating that the Ctrl-CLSTM is capable of accurately modeling these different reservoirs. From Table 4, we can conclude that the value of SSIM is over 96% and the MSE is around 21. These data are very close to those derived from our experiments done on the real Tarim dataset above, proving the compatibility of the Ctrl-CLSTM with different reservoirs.

6. Conclusions

As an important application of deep learning, generating future sequence from historical continuous sequence has become a hot topic of interest. In this paper, we proposed a novel unsupervised network named Ctrl-CLSTM for spatiotemporal predictive learning tasks, which realized the unified modeling of the spatiotemporal deformations simultaneously. The core part of the Ctrl-CLSTM was the dual-memory state transmit mechanism that made temporal features propagate horizontally over time and spatial features span stacked network layers vertically. Furthermore, we innovatively introduced a control gate structure in the spatiotemporal sequence prediction network to cope with a special sequence such as oil saturation. The Ctrl-CLSTM took the existing oil saturation sequence, well location maps and injection–production parameters as input to predict the future oil saturation. When changing the injection–production parameters, the Ctrl-CLSTM can obtain different prediction results, which can save the oilfield prediction time and help to enhance the oil recovery.
In order to capture the trend of oil saturation more accurately, the Ctrl-CLSTM requires the historical data to be used in the prediction; therefore, it is mainly used in the secondary oil recovery stage to adjust the injection–production parameters continuously, but it cannot be applicable to new reservoirs that are completely undeveloped. Furthermore, when using the Ctrl-CLSTM to generate oil saturation, we did not consider sufficient factors that affect the variation of oil saturation, and in the future, we will consider more influencing factors and adapt it to more application scenarios.

Author Contributions

Conceptualization, F.L. and Y.D.; methodology, F.L.; software, F.L.; validation, Y.Z. and F.L.; formal analysis, F.L.; investigation, F.L.; resources, Q.W.; data curation, Q.W.; writing—original draft preparation, F.L.; writing—review and editing, F.L.; visualization, Y.Z.; supervision, Y.D.; project administration, Y.D.; funding acquisition, Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fundamental Research Funds for the Central Universities (20CX05016A) and the Major Scientific and Technological Projects of CNPC under Grant ZD2019-183-007.

Data Availability Statement

The data are not publicly available for the following reasons: for the data provided by Tarim, we only use the data for the prediction of oil saturation in the Tarim project, and not for any other projects or purposes.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Z.; Zhang, G.; Jin, Y. Quantitative characterization of oil recovery and pore volume of water injected in water flooding. J. China Univ. Pet. 2021, 45, 94–100. [Google Scholar]
  2. Zhong, Z.; Sun, A.Y.; Wang, Y.; Ren, B. Predicting field production rates for waterflooding using a machine learning-based proxy model. J. Pet. Sci. Eng. 2020, 194, 107574. [Google Scholar] [CrossRef]
  3. Zhang, K.; Chen, G.; Xue, X.; Zhang, L.; Sun, H.; Yao, C. A reservoir production optimization method based on principal component analysis and surrogate model. J. China Univ. Pet. 2020, 44, 90–97. [Google Scholar]
  4. Yang, Y.; Yang, H.; Tao, L.; Yao, J.; Wang, W.; Zhang, K.; Luquot, L. Microscopic determination of remaining oil distribution in sandstones with different permeability scales using computed tomography scanning. J. Energy Resour. Technol. 2019, 141, 092903. [Google Scholar] [CrossRef]
  5. Liu, P.; Zhang, Z.; Wang, X.; Pan, M.; Shi, Y. Revelation on remaining oil and full length hydraulic simulation of single sand bodies based on 3D models in low permeability heterogeneous reservoirs. J. China Univ. Pet. 2017, 41, 56–64. [Google Scholar]
  6. Liu, Y.; Hou, S.; Wang, Z.; Yu, X.; Li, S. Reservoir prediction and application based on AVO three parameter seismic inversion under condition of dense well pattern. Fault-Block Oil Gas Field 2020, 27, 750–753. [Google Scholar]
  7. Liu, T.; Li, C.; Liu, C.; Sun, Q. Experimental simulation of remaining oil distribution in combined debouch bar of delta front reservoir. J. China Univ. Pet. 2018, 42, 1–8. [Google Scholar]
  8. Gu, J.; Ren, Y.; Wang, Y. Prediction methods of remaining oil plane distribution based on machine learning. J. China Univ. Pet. 2020, 44, 39–46. [Google Scholar]
  9. Tang, M.; Liu, Y.; Durlofsky, L.J. A deep-learning-based surrogate model for data assimilation in dynamic subsurface flow problems. J. Comput. Phys. 2020, 413, 109456. [Google Scholar] [CrossRef] [Green Version]
  10. Shi, X.; Yeung, D.Y. Machine learning for spatiotemporal sequence forecasting: A survey. arXiv 2018, arXiv:1808.06865. [Google Scholar]
  11. Zhang, K.; Wang, Y.; Li, G.; Ma, X.; Cui, S.; Luo, Q.; Wang, J.; Yang, Y.; Yao, J. Prediction of Field Saturations Using a Fully Convolutional Network Surrogate. SPE J. 2021, 26, 1824–1836. [Google Scholar] [CrossRef]
  12. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.c. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In Proceedings of the Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  13. Meng, F.; Song, T.; Xu, D.; Xie, P.; Li, Y. Forecasting tropical cyclones wave height using bidirectional gated recurrent unit. Ocean Eng. 2021, 234, 108795. [Google Scholar] [CrossRef]
  14. Shi, X.; Gao, Z.; Lausen, L.; Wang, H.; Yeung, D.Y.; Wong, W.k.; Woo, W.c. Deep learning for precipitation nowcasting: A benchmark and a new model. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  15. Wang, Y.; Long, M.; Wang, J.; Gao, Z.; Yu, P.S. Predrnn: Recurrent neural networks for predictive learning using spatiotemporal lstms. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  16. Wang, Y.; Wu, H.; Zhang, J.; Gao, Z.; Wang, J.; Yu, P.S.; Long, M. PredRNN: A recurrent neural network for spatiotemporal predictive learning. arXiv 2021, arXiv:2103.09504. [Google Scholar] [CrossRef]
  17. Feng, Q.; Li, Y.; Wang, S. Predicting gas migration development using deep convolutional generative adversarial network. J. China Univ. Pet. 2020, 44, 20–27. [Google Scholar]
  18. Hu, Y.; O’Donncha, F.; Palmes, P.; Burke, M.; Filgueira, R.; Grant, J. A spatio-temporal LSTM model to forecast across multiple temporal and spatial scales. arXiv 2021, arXiv:2108.11875. [Google Scholar]
  19. Tang, M.; Liu, Y.; Durlofsky, L.J. Deep-learning-based surrogate flow modeling and geological parameterization for data assimilation in 3D subsurface flow. Comput. Methods Appl. Mech. Eng. 2021, 376, 113636. [Google Scholar] [CrossRef]
  20. Gnewuch, M.; Hebbinghaus, N. Discrepancy bounds for a class of negatively dependent random points including Latin hypercube samples. Ann. Appl. Probab. 2021, 31, 1944–1965. [Google Scholar] [CrossRef]
  21. Dong, Y.; Zhang, Y.; Liu, F.; Cheng, X. Reservoir Production Prediction Model Based on a Stacked LSTM Network and Transfer Learning. ACS Omega 2021, 6, 34700–34711. [Google Scholar] [CrossRef]
  22. Sukhbaatar, S.; Weston, J.; Fergus, R. End-to-end memory networks. In Proceedings of the Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  23. Wu, J.; Li, Z.; Sun, Y.; Cao, X. Neural network-based prediction of remaining oil distribution and optimization of injection-production parameters. Pet. Geol. Recovery Effic. 2020, 27, 85–93. [Google Scholar]
  24. Yang, X. Waterflood Development Effect Study of Heterogeneous Reservoir. Spec. Oil Gas Reserv. 2019, 26, 152–156. [Google Scholar]
  25. Dağdeviren, A.; Kürüz, F. Special Real and Dual Matrices with Hadamard Product. J. Eng. Technol. Appl. Sci. 2021, 6, 127–134. [Google Scholar] [CrossRef]
  26. Bock, S.; Weiß, M. A proof of local convergence for the Adam optimizer. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
  27. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  29. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
Figure 1. Architecture of Ctrl-CLSTM.
Figure 1. Architecture of Ctrl-CLSTM.
Energies 15 05063 g001
Figure 2. Units of Ctrl-CLSTM.
Figure 2. Units of Ctrl-CLSTM.
Energies 15 05063 g002
Figure 3. Permeability distribution and well location.
Figure 3. Permeability distribution and well location.
Energies 15 05063 g003
Figure 4. Input sequence and numerical simulation sequence.
Figure 4. Input sequence and numerical simulation sequence.
Energies 15 05063 g004
Figure 5. Prediction examples on different models.
Figure 5. Prediction examples on different models.
Energies 15 05063 g005
Figure 6. Prediction effects over time.
Figure 6. Prediction effects over time.
Energies 15 05063 g006
Figure 7. Prediction examples on different injection–production working systems.
Figure 7. Prediction examples on different injection–production working systems.
Energies 15 05063 g007
Figure 8. Prediction examples.
Figure 8. Prediction examples.
Energies 15 05063 g008
Table 1. Partial injection–production parameters.
Table 1. Partial injection–production parameters.
T 1 G 1 T 4 G 4 W 1 W 12 Q 1 Q 21
3.593,459924911241205437
0113,668836791437810170
244,748368152168123117
6.510450.5101,05727768533
862,0629.576,50567407485
1.558,7847.529,22052317030
025,8077.5116,0921261225428
4.533,2295114,7201321447880
Table 2. Quantitative results on each evaluation metric.
Table 2. Quantitative results on each evaluation metric.
ModelSSIM (%)MSELPIPSPSNR (dB)
FC-LSTM83.1636.6111.2932.06
TrajGRU89.1033.7810.0436.41
ST-LSTM93.5627.194.2941.73
Ctrl-CLSTM98.2321.372.4146.23
Table 3. Quantitative results with different layers.
Table 3. Quantitative results with different layers.
Model LayersSSIM (%)MSELPIPSPSNR (dB)
Two layers90.1327.826.2442.08
Three layers98.2321.372.4146.23
Four layers92.0725.618.9339.75
Table 4. Quantitative results on two sequences.
Table 4. Quantitative results on two sequences.
Input SequenceSSIM (%)MSELPIPSPSNR (dB)
(a)96.5920.322.7644.91
(b)97.2721.072.5945.26
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dong, Y.; Liu, F.; Zhang, Y.; Wu, Q. Prediction of Oil Saturation during Water and Gas Injection Using Controllable Convolutional Long Short-Term Memory. Energies 2022, 15, 5063. https://doi.org/10.3390/en15145063

AMA Style

Dong Y, Liu F, Zhang Y, Wu Q. Prediction of Oil Saturation during Water and Gas Injection Using Controllable Convolutional Long Short-Term Memory. Energies. 2022; 15(14):5063. https://doi.org/10.3390/en15145063

Chicago/Turabian Style

Dong, Yukun, Fubin Liu, Yu Zhang, and Qiong Wu. 2022. "Prediction of Oil Saturation during Water and Gas Injection Using Controllable Convolutional Long Short-Term Memory" Energies 15, no. 14: 5063. https://doi.org/10.3390/en15145063

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop