Transformer with Sparse Mixture of Experts for Time-Series Data Prediction in Industrial IoT Systems ()
1. Introduction
Oil and gas industry operations from mining and extraction to transportation require efficient and reliable techniques to handle critical situations in oil fields for more effective recovery and higher yields. For example, a wellhead, which is requisite for the reliability and efficiency of oil and gas wells, is used to manage the extraction of hydrocarbons from underground formations and prevent pressure-induced oil or gas leakages from the well. Real-time monitoring of the oil wellhead production status is crucial for safe operation and productivity and for prolonging the production life of the well. Owing to the vast distances between oil wells and their remote locations, it is not feasible to manually monitor all the production data in real time, as such an undertaking would be labor-intensive, inefficient, and time-consuming. Moreover, the conventional manual method of onsite data acquisition is inevitably subject to human errors and data forgery [1] [2]. Thus, forecasting oil well production has attracted increasing attention in the oil and gas industry to support petroleum engineers in analyzing the features of crude oil wells. However, accurately forecasting oil and gas production can be challenging and requires a large amount of data and advanced technologies [3].
The Internet of Things (IoT) is a result of the convergence of various technologies, such as wireless communication and smart sensors [4] [5]. This convergence has led to the development of the industrial IoT (IIoT), which is a promising technology that can significantly reduce costs and increase efficiency. IIoT technology provides a good alternative for monitoring and controlling various processes and operations, and it has been widely used in the oil and gas industry. An online IoT-based real-time monitoring system contributes to the timely acquisition of accurate production parameters, such as oil wellhead status data, critical oil-well production data, and data related to onsite operational activities [6] [7], while minimizing the safety risks in the oil and gas industry [8]-[10]. The vast amounts of time-series data continuously generated by smart sensors are essential for the real-time monitoring of production, as well as for intelligent analysis and decision-making [11] [12]. For example, the future trend can be predicted through the periodic pattern of the data, and sudden changes in time-series data suggest anomalies in the actual production line [13]. Moreover, the loss of temporal data due to the absence of critical metering equipment, sensor failure, or network congestion can be detected, and the data can be replaced.
Deep learning algorithms are innovative data analysis techniques that utilize advanced neural networks to analyze and predict data according to large volumes of correlated time-series data. A typical method for increasing the accuracy and efficiency of predictions is to process real-time and historical data for relational trend prediction and missing data complementation [14]-[17]. From the perspective of deep learning algorithms, various Transformer-based models have been proposed for long sequence prediction. These models effectively reduce the complexity of time and space by incorporating a sparse attention mechanism [18] [19]. Lepikhin et al. introduced a Transformer model with conditional computations by replacing every other feed forward layer with a MoE layer to enhance the model capacity [20]. Zoph et al. addressed training process instabilities by scaling a sparse MoE layer [21]. These investigations have illustrated the potential of the Transformer model in data analysis, owing to its ability to effectively capture the extensive interrelationships between the multiple input variables [22] [23]. However, few reports exist on the use of a Transformer model with MoE block (referred to as TransMoE) for the missing data trend prediction in multivariate data.
This study proposes a TransMoE algorithm designed for a cloud-assisted IIoT system. The algorithm utilizes correlated multivariate time-series datasets to forecast petroleum production and replaces missing critical production data. This is the first hybrid TransMoE algorithm to be proposed for predicting oil well production using original time-series data and historical data sequences. The existing deep-learning algorithms based on a CNN along, gated recurrent unit (GRU), or long short-term memory (LSTM) model can only extract local features from the original data. In contrast, the TransMoE algorithm is capable of extracting long-range dependencies from raw historical data without extensive prior knowledge. It can reveal data associations over time and accurately predict future data trends without complete knowledge of the raw data. The prediction accuracy of the proposed model was evaluated using a real industrial dataset. The results indicated that the proposed model achieved time-series data prediction with an average relative error of 6.26%.
The remainder of this paper is organized as follows. Section 2 reviews previous research on time-series data prediction. Section 3 presents the cloud-assisted IIoT system. In Section 4, the proposed hybrid TransMoE model and its application to missing-data prediction based on historical data are described. The performance evaluation of the proposed model is presented in Section 5. Finally, Section 6 provides a summary of the paper and presents conclusions.
2. Related Work
This section reviews previous research that is relevant to our study. IIoT has been widely used in various applications, primarily because of the increased frequency and efficiency of data collected from the field perception layer, which allows the status of specific equipment to be monitored over time. Identifying missing data or unusual conditions monitored by sensors is often necessary. This process is designed to detect defective equipment, detect quality problems, and alert supervisors to abnormal changes [2]. Because of the massive and real-time nature of time-series data, using historical data to predict future trends and complete missing data requires a large amount of computing and storage resources and must be implemented in the cloud. Key challenges in the oil industry include integrating in situ data collected by intelligent sensors on remote terminal units (RTUs) and real-time analytics in the cloud [15] [17]. The heterogeneous sensed data are generated by different sensors and data communication protocols.
Researchers have studied the integration of heterogeneous data, i.e., the combination of data produced by various sensors or devices. Additionally, researchers have investigated various deep-learning algorithms for discovering hidden patterns in time-series data, predicting future trends, and completing missing data using historical data. Deep learning has considerable potential as a comprehensive data-driven analytics method for making predictions and learning temporal contextual information [20] [21]. Researchers have successfully applied neural networks in the oil and gas industry to forecast oil well production and predict time-series data [3] [14] [22].
Deep learning methods excel at learning useful representations of multidimensional data features and their nonlinear interactions from high-dimensional raw data. However, multiple linear regression algorithms [23], e.g., the CNN-LSTM [24], CNN [25], and GRU [26] models, can only focus on the short- and medium-term characteristics of the original sequence. The simulation performance for the correlation characteristics of high-dimensional data has not been fully validated. Variations in the production of a single oil well are influenced by many factors, such as reservoir properties and stimulation measures. To achieve accurate production forecasting, it is necessary to consider each element comprehensively and thoroughly and to fully exploit the data features that affect production variability. Compared with traditional production prediction methods, deep-learning methods have a solid nonlinear fitting capability and higher computational efficiency, giving them considerable potential for application in production prediction. For single-well production prediction with many influencing factors, long periods, and parallel sequences. However, when the number of features increases, it is difficult to extract high-dimensional spatial and temporal information, and the prediction accuracy is limited.
Recently, self-attention mechanisms have exhibited remarkable capabilities in sequential data, such as natural language processing and computer vision. In a supply chain, a slight modification to the original transformer was used to forecast customer sales at the day, store, and item levels according to time-series data [27]. LogTrans introduces convolutional self-attention by generating queries and keys with causal convolution and proposes sparse attention to incorporate the local context into the attention mechanism [28]. Informer extends Transformer with ProbSparse attention by halving the cascading-layer input and efficiently handles extremely long input sequence predictions [17]. Lepikhin et al. introduced a Transformer model with conditional computations by re-placing every other feed forward layer with a MoE layer to enhance the model capacity [18]. Zoph et al. addressed training process instabilities by scaling a sparse MoE layer [19]. These investigations have illustrated the potential of the Transformer deep learning algorithm in data analysis, owing to its ability to effectively capture the extensive interrelationships between the multiple input variables [20] [29].
In this paper, a new oil well yield prediction algorithm based on the TransMoE model using the self-attention mechanism is proposed. When the standard attention mechanism for time-series data is combined with the MoE, it relies heavily on the multitasking features of the MoE operations. Multivariate trends in the raw data are extracted by using different expert processing submodules. This results in additional features when calculating attention scores that can reflect the short- and long-range dependencies of the input data sequences in the Transformer part. The proposed TransMoE algorithm can be used for sequential data prediction in cloud-assisted IIoT systems. In the TransMoE model, a multihead attention mechanism is used as a sequence encoder to calculate attention scores to obtain a more complete temporal dependence compared with standard self-attention. To evaluate the effectiveness of our model, it was compared with the CNN-GRU and CNN-LSTM models. Extensive empirical results for a real industry dataset indicated that the model outperformed the existing methods.
3. Cloud-Assisted IIoT System
In cloud-based IIoT applications, the high heterogeneity of sensed data from various intelligent sensors and the fusion between other measured data should be considered. Research on real-time data analysis for oil and gas wells requires multifunctional intelligent sensor nodes that can withstand harsh environments, which inevitably produce missing data.
Figure 1. Architecture of the cloud-assisted IIoT system for the oil and gas industry.
Therefore, we present a system in which data are obtained from oil wells through an integrated RTU compatible with heterogeneous smart sensors and communication protocols. A deep-learning method is implemented in the application layer to detect missing sensory data, predict and complete the missing data. The developed IoT system is useful for isolated industrial applications in remote desert areas. An overview of the four-tier cloud-assisted IIoT system is presented in Figure 1. Data collected from sensors are sent to the RTU only when the request message is received. The network layer bridges the perception layer and the server resource layer. The TransMoE algorithm is embedded in the application layer for future data trend prediction and completion of missing data.
3.1. Smart Perception and Network Layer
The smart perception layer is primarily used for data collection and consists of various sensors for different parameters, as shown in Table 1. Pressure and temperature are monitored wirelessly, whereas other measurements are performed using wired instruments. The diverse data collected from various sensors are transmitted through different communication protocols in the physical layer. The network layer performs the essential functions of data transmission. A sink node is used in the network layer as a gateway that transmits the protocol packet to the remote cloud server in a transparent manner. As shown in Table 1, in the oil industry, typical heterogeneous data include pressure, temperature, and oil-well production data collected at the wellhead, along with voltage and current data from pumps.
Table 1. Monitoring data items.
Parameters |
Units |
Parameters |
Units |
Oil-well production |
tons |
Displacement of pump |
m3 |
Oil tubing pressure |
MPa |
Stroke of pump |
m |
Casing pressure |
MPa |
Frequency of pump |
min |
Pipeline pressure |
MPa |
Voltage |
V |
Oil wellhead temperature |
˚C |
Current (up) |
A |
Liquid production |
tons |
Current (down) |
A |
Gas production |
m3 |
Electricity consumption |
kWh |
Depth of pump |
m |
Weight of blend liquid |
tons |
3.2. Service Resource and Application Layer
The service resource layer composed of different servers is deployed to process and analyze sensed data. It includes a data acquisition server, a real-time database server, a relational database server, and an application server. Each data acquisition server receives real-time data from hundreds of oil wells and dump data into the relational database. Moreover, hourly data are stored in the relational database, which is used for data queries in applications with low real-time requirements. The application server performs real-time data processing and high-performance computing, which uses deep-learning algorithms to process, analyze, and compute data for trend prediction and automatic replenishment of abnormally missing data.
4. Proposed Transformer Model with Sparse Mixture of Experts
The proposed TransMoE model fuses a MoE block and a Transformer with a self-attention mechanism for time-series data prediction. The Transformer model based on the encoder-decoder architecture includes a multi-head attention module and feedforward neural network layer. A residual connection and layer normalization are used in the encoder and decoder sections to prevent model degradation, as shown in Figure 2. The encoder generates the vector corresponding to the input sequence, and the decoder generates the target sequence according to the output of the encoder. The encoder maps the input sequence
with n features to the linear embedding sequence data
and then feeds it to a multi-head attention layer. The decoder generates the output sequence
.
Figure 2. Architecture of the Transformer with a sparse MoE block (TransMoE).
CNN has good performance in local feature extraction, which can affect the accuracy of the predicted values. The convolution operation primarily uses feature-extraction blocks consisting of three Conv1D layers. To extract the local features of the input data, three layers of 1D convolution operations are performed on the original input data before inputting the embedding layer to enhance the data representation, with each layer having 128 neurons and a convolution kernel size of 11. However, they fail to capture long-range dependencies compared with the Transformer model and require deeper networks with multiple layers to increase their receptive fields. Combining the efficiency and inductive prior of CNNs with the ability to capture information over long distances can yield effective architectures for sequential data applications [30] [31]. To capture the long-range dependencies of the input sequential data, the proposed model is adapted to include a sparse MoE layer to enhance the input sequence data.
The positions of the input data sequence constitute valuable information. To make the multihead attention layer aware of the sequence order, information about the relative positions of the tokens in the sequence must be injected. The fixed sine and cosine functions’ positional encoding is used to identify the position information. The position encoding was implemented as follows:
(1)
where
represents the position of the embedding tokens,
is the embedding depth index in the range of
, and
represents the embedding depth. The values generated by the sine and cosine functions are concatenated pairwise and added to the embedding of the input sequence.
The multihead attention layer is the main part where attention scores are calculated, and the attention scores of the input sequence are modeled using a self-attention mechanism based on three main concepts: the query vector
, the key vector
, and the value vector
. These three pieces represent an analogy to information-retrieval systems where a query is used to search for the matching key (or the most similar one) and retrieve its value. There are many different similarity functions that can be used. The scaled dot product is a similarity function, given that it is scaled so that sequences of different lengths can be easily compared. A single sequence query searches potential relationships by finding similarities in the sequence through keys. Comparing the query and key pairs gives attention weight to the value. The interaction between the attention weights and values determines how much focus to place on other parts of the sequence while representing the current sequential data. The query, key, and value matrices are calculated by multiplying the input sequence
by three different weight matrices:
,
, and
.
(2)
In Equation (2),
,
, and
represent the query, key, and value, respectively. The multi-head attention calculations are implemented via scaled dot-product attention, as shown in Figure 2. Scaled dot-product attention is expressed by Equation (3). The weight is calculated through the query and key, and it is used to obtain a weighted sum of values.
(3)
here,
represents the dimension of the key vector sequence. The ith attention head can be calculated as follows:
(4)
where
,
, and
denote the linear transformations of
,
, and
, respectively, of the ith attention head. The parametric matrices
,
,
, and
are linear projection parameters, and
represents the number of heads in the multi-head attention.
and
represent the dimensions of the key and query, respectively. Concat denotes the concatenation operation, and
denotes the linear transformation of concatenated outputs. The multi-head attention model can focus on the different representation spaces and simulate multiple levels of detailed information.
The feedforward layer comprises two linear transformations and a rectified linear unit activation function. The computation of the feedforward network is positionwise, and the weights of the linear transformations of the different steps are identical. The formula of the feedforward network is as follows:
(5)
The last two layers in the decoder are an affine transformation layer, which consists of a learnable scaling factor and a bias factor, and a multilayer perceptron, which consists of two linear transformations. The affine transformation layer and the multilayer perceptron are sequentially used for the trend prediction of the sequential data. The corresponding formulas are as follows:
(6)
where
,
,
, and
are linearly learnable weights and biases, respectively.
The last layer in the decoder is a sparse MoE block, which contains a set of
expert networks,
, and a gating network,
whose output is a sparse
-dimensional vector. The experts are neural networks in which the experts accept the same-sized inputs and produce the same-sized outputs, each with their own parameters. For a given input
, the output of the gating network is denoted as
, and the output of the
-th expert network is denoted as
. The output
of the MoE module can be written as follows in (7) and (8).
(7)
(8)
The Softmax gating network has two components: sparsity and noise. Before taking the Softmax function, we add tunable Gaussian noise and then retain only the top
values, setting the rest to minus infinity. Sparsity serves to save computations, while this form of sparsity creates discontinuities in the output of the gating function. The noise term helps with load balancing.
5. Evaluation and Discussion
We evaluated the performance of missing-data prediction based on the TransMoE model and the generated dataset. The prediction results were compared with those of the CNN-GRU and CNN-LSTM models and the actual data to evaluate their accuracy. The two baseline prediction models, i.e., CNN-GRU and CNN-LSTM, included convolutional blocks, each consisting of a 3D convolutional layer with 128 hidden units and a kernel size of 11. In the comparison experiments, the models utilized the same time-series oil wellhead production data from the oil and gas industry.
5.1. Hybrid CNN-GRU Model
We developed a deep-learning model incorporating a multilayer CNN-GRU model and used it to evaluate the missing-data prediction performance of the TransMoE model [32] [33]. The architecture of the CNN-GRU model is shown in Figure 3, which consists of two main neural network models. With regard to the structure and parameters, the GRU is superior to the LSTM neural network. It can optimize the computation of hidden states in recurrent neural networks, where the reset gate controls how much information from the previous state is written to the current candidate set and discards historical information irrelevant to the prediction. The update gate can control the degree to which the state information of the previous moment is brought into the current state.
Figure 3. Proposed CNN-GRU model-based framework for evaluation of the prediction.
The GRU is used to extract the data characteristics of time-series data, and its architecture consists of a series of neuronal cells, each containing two gates, a reset gate
and an update gate
and one hidden state layer. The hidden state layer is a crucial variable that carries information from the previous step. The gate cells in the interaction layer can partially remove the state of the previous step and add new information to the current step according to the hidden state of the previous step and the input of the current step. The corresponding formula is as follows:
(9)
(10)
In Equation (9),
is a sigmoid function given by Equation (10).
denotes the feature vector at timestep
.
denotes the previous state at time
.
, and
and
, and
are weight matrices and deviation vectors, respectively, which contribute to the linear transformations of
and
. The update gate
adds this linearly transformed information and inputs it to the sigmoid activation function, which can compress the result between 0 and 1. The other gate (
) is the reset gate, which is used to determine how much the candidate state at the current moment will inherit from the previous moment.
denotes element multiplication,
denotes the hidden state, and
denotes the candidate hidden state.
5.2. Dataset
To evaluate the efficiency and accuracy of the proposed TransMoE model for time-series data prediction, we selected 16 critical parameters from oil wells located in the most remote areas for testbed scenarios. Sensor nodes collected these data at different points in the perception layer in the cloud-assisted IIoT system. The dataset was recorded with a 5-min resolution over approximately 45 d. It included 16 parameters related to pressure, temperature, pump information, liquid production, oil-well production, and voltage values. All 16 parameters were normalized with respect to the minimum and maximum values to generate the graph shown in Figure 4. The bottom half of this figure presents a comprehensive graph of the last data series of length 256. Consequently, the collected data included a large number of records. However, most of the nodes in the dataset had numerous missing or corrupted records. Figure 4 shows the profiles of the selected parameters over a span of 45 d.
As mentioned in Section 2, oil well yields are closely related to the production parameters, such as the pressure, temperature, pump information, liquid production, and voltage values. Accordingly, the input data included 15 parameters, such as pressure, temperature, and pump information, and the output data were the oil-well yields. The dataset was divided into a training set, validation set, and test set according to the input and output sequences. One stride of raw data was used to generate a series of sequential datasets of 256 data points; thus, 2722 slices of data were created, with final data dimensions of (2722, 256, 16). A three-dimensional (3D) tensor was used to store the oil well production data with dimensions of (sample, sequence, and features). Here, “sample” refers to the sample size of the dataset, which contained 2722 observations. Because the objective of this study was to forecast the oil well production for each oil well over a certain period, we set the sequence length to 256.
To predict the oil well yields, we resampled the real-time data and processed them into daily data. Thus, we selected a data sequence without missing data to train the model for trend prediction and missing-data complementation. Details regarding the selected data are presented in Table 1. To increase the correction accuracy of the measured data, a large convolution kernel of size 11 was applied to the original data sequentially to extract the local features. The input data sequence was first used as the input to three consecutive convolutional layers. Therefore, the low-level features and distinctions among variables under the context of temporal effects were acquired through the convolution operations of filters with different properties and nonlinear activation of neurons. The obtained feature map was then passed to the transformer or GRU layer, where the multihead attention mechanism thoroughly learns the complex long-range dependencies or the examination of three effective reset and update gates. Finally, the information was projected to the output space, and the prediction results were produced.
![]()
![]()
Figure 4. Raw data curves for the 16 selected parameters of the dataset.
5.2. Model Training
The proposed deep learning model using the TransMoE algorithm was implemented using Python 3.7 and PyTorch 1.9.1 and was run on a Tesla P100 GPU server. The data were split into three sets for training, validation, and testing. These datasets were normalized with respect to the minimum and maximum values in the range of 0 - 1. The first 2256 data sequences served as the training set, the intermediate 210 sequences served as the validation set, and the final sequence served as the test set. A blank 256 data sequence length between the validation and test sets ensured that it was not used in the training and testing process. The batch size was set as 50, with a total of 300 training epochs. For each training epoch, approximately 45 groups of training data were randomly selected. To achieve optimal performance, the parameters were tuned for each prediction model.
5.3. Results and Discussion
To evaluate the proposed model, three deep-learning models, i.e., CNN-GRU, CNN-LSTM, and TransMoE, were employed for predicting oil-well production over a span of time. One sequence of time-series data including 256 samples (several oil well time-series data) was selected from the test dataset. The prediction results are presented in Figure 6, where the red, brown, and light green curves indicate the oil-well yields predicted by TransMoE, CNN-LSTM, and CNN-GRU, respectively. As shown, all the deep learning models can predict future trends. Moreover, the prediction results are beneficial for solving the problem of missing data due to device failure or network interruptions. The prediction performance was assessed using the bias, mean absolute error (MAE), and mean absolute percentage error (MAPE) metrics defined in Equation (11). The MAE measured the absolute difference between the actual and predicted values for the deep-learning models.
(11)
here,
represents the actual value,
represents the predicted value, and
represents the total number of samples. The MAPE indicator of the error metrics was introduced because different models have arbitrarily different error levels. The usage of MAPE normalizes the error distribution and makes the whole prediction easier to measure.
Table 2. Performance of the three deep learning models.
Model |
Minimum bias |
Maximum bias |
MAE |
MAPE |
Computation time (s) |
TransMoE |
0.03 |
1.96 |
0.84 |
6.26% |
25867 |
CNN-GRU |
−1.39 |
2.17 |
0.85 |
7.00% |
6135 |
CNN-LSTM |
−2.39 |
3.09 |
0.50 |
6.84% |
9212 |
Compared with the other two models, the TransMoE model exhibited better performance with regard to the curve morphological characteristics of the predicted values and the calculated bias, MAE, and MAPE values. For instance, the minimum and maximum biases, MAE, and MAPE of the TransMoE model were 0.03, 1.96, 0.84, and 6.26%, respectively, which were better than those of CNN-GRU and CNN-LSTM, as shown in Table 2. In particular, the CNN-LSTM model had larger minimum and maximum deviations, resulting in larger deviations in the predicted values at local points.
For the CNN-GRU and CNN-LSTM models, the minimum and maximum biases resulted in poor performance, which may have been related to the network structure. These models had three CNN layers and three GRU layers. The hidden neurons in the CNN layer extract local features from the input sequence and feed them to GRU or LSTM layers, which store information from the previous step. The hidden nodes in the GRU or LSTM layer transmit information forward, and it eventually reaches the output layer.
Figure 5 shows the distribution of bias errors for eight different oil wells. The errors for the different wells are shown side-by-side to illustrate how they vary across the oil wells. As indicated by the results, the TransMoE model had the smallest bias variation range, along with the smallest local minimum and maximum values. This implies that the TransMoE model achieved the best fit to the actual data and most accurately forecasted the data’s long-term trend. Although the CNN-GRU and CNN-LSTM models accurately approximated the actual values in certain periods, there were large deviations in the predicted values between different oil well data variations. Moreover, the bias in the predicted values was large, suggesting that these two models can simulate the short-term characteristics of the data but have difficulty accurately predicting long-term trends because of the limitations of the algorithms.
![]()
Figure 5. Distribution of prediction errors at different times for the three models and for the eight different wells.
Figure 6 presents a comparison of the actual values and the predicted values of the three deep learning models. As shown, the TransMoE model had the highest prediction accuracy among the models. Comparing the three models horizontally and vertically revealed that the prediction accuracies of the TransMoE model were higher than those of the CNN-LSTM and CNN-GRU models.
Figure 6. Predicted values of the three deep-learning models and the actual values.
A visual analysis of the bias between the predicted oil well production data and the actual data revealed that the proposed model can accurately predict oil well production levels for different wells. The predicted data changed in accordance with the actual oil well production data in a predicted data series, and the local maximum and minimum values were perfectly fitted. Larger variations in the actual data corresponded to a higher accuracy of the predicted trend. This indicates the capability of the TransMoE model for capturing the long-term dependencies of time-series data. In comparison, CNN-GRU and CNN-LSTM were slightly inferior in fitting the long-term trend patterns of the sequential data.
The findings indicate that the proposed algorithm is promising for solving time-series problems—particularly in long-term forecasting tasks. This is because it accurately characterizes long-interval time-series datasets better than conventional methods. The proposed model is useful for the oil and gas industry because of its ability to handle local features as well as long-term dependencies caused by the characteristics of time-series data, despite having a higher computational cost than CNN-GRU and CNN-LSTM.
Our simulation results indicated that the CNN-GRU model performed at a comparable level to the CNN-LSTM model for the test set. The gated neural network model can only focus on the short-term characteristics of the data and cannot learn the long-term dependencies of the data, hindering the algorithm’s ability to predict long-term trends.
6. Conclusion
Forecasting production for a single well can be challenging, as oil production lacks consistency and tends to fluctuate even on consecutive days. We developed a TransMoE model for predicting oil well yields in the oil and gas industry. The model was tested using real industrial datasets in an IIoT system. As the input data were nonlinear, they were normalized using a standard min-max scalar before being fed into various training processes. Several deep learning models were investigated, and the Transformer deep learning model with a sparse MoE block was optimized. The proposed model outperformed other deep-learning models, indicating that it can be used in practical applications to predict time-series data. To assess the prediction accuracy, the differences between the predicted and actual data were evaluated using the bias, MAE, and MRE. The experimental results indicated that deep learning algorithms can achieve high prediction accuracy and can be used in practical applications.
Acknowledgements
We would like to thank Editage (http://www.editage.cn/) for English language editing.