- 1China University of Petroleum (East China), Qingdao, China
- 2Hebei University of Science and Technology, Shijiazhuang, Hebei, China
Unconventional reservoirs are rich in petroleum resources. Reservoir fluid property identification for these reservoirs is an essential process in unconventional oil reservoir evaluation methods, which is significant for enhancing the reservoir recovery ratio and economic efficiency. However, due to the mutual interference of several factors, identifying the properties of oil and water using traditional reservoir fluid identification methods or a single predictive model for unconventional oil reservoirs is inadequate in accuracy. In this paper, we propose a new ensemble learning model that combines 12 base learners using the multiverse optimizer to improve the accuracy of reservoir fluid identification for unconventional reservoirs. The experimental results show that the overall classification accuracy of the adaptive ensemble learning by opposite multiverse optimizer (AIL-OMO) is 0.85. Compared with six conventional reservoir fluid identification models, AIL-OMO achieved high accuracy on classifying dry layers, oil–water layers, and oil layers, with accuracy rates of 94.33%, 90.46%, and 90.66%. For each model, the identification of the water layer is not accurate enough, which may be due to the classification confusion caused by noise interference in the logging curves of the water layer in unconventional reservoirs.
1 Introduction
Reservoir fluid property identification is an important method for evaluating oil reservoirs as it plays a key role in calculating oil reserve production and formulating oil field development plans. Unconventional oil reservoirs, with their low permeability and nonhomogeneous and complex pore structure, present challenges in identifying reservoir fluid properties. Currently, reservoir fluid identification relies on various logging data to explore the intrinsic connections among logging data, geological data, and oil and gas water information in the reservoir. There are four types of methods used in reservoir fluid identification: petrophysical methods (Das and Chatterjee, 2018), cross-plot techniques based on cores or log responses, experimental formula methods (Moffatt and Williams, 1998), and artificial intelligence methods (Bestagini et al., 2017). Due to the mutual interference of many factors, such as reservoir lithology and pore structure in unconventional oil reservoirs, identifying the properties of oil and water using traditional reservoir fluid identification methods remains a challenge. Therefore, artificial intelligence techniques have been applied in this field, providing new perspectives on fluid identification in reservoirs.
With the rapid development of artificial intelligence techniques in recent years, regression algorithm (Bestagini et al., 2017), support vector machine techniques (Al-Anazi and Gates, 2010; Tohidi-Hosseini et al., 2016), clustering algorithm (Baarimah et al., 2019), genetic algorithm (Guerreiro et al., 1998), artificial neural network (Onwuchekwa, 2018), decision tree algorithm (He et al., 2020), random forest (Wang et al., 2020), and thermodynamics-informed neural network (Zhang and Sun, 2021) are used in oil reservoir evaluation.
In reservoir fluid identification, Yang et al. (2016) used a synergistic wavelet transform and improved K-means clustering technology to classify reservoir fluid. Al-Anazi and Gates (2010) applied an SVM model to nonhomogeneous sandstone reservoirs to classify reservoir fluid. The results obtained by Anazi demonstrated that SVM is superior to traditional models in data training and model generalizability of reservoir fluid in nonhomogeneous sandstone reservoirs. Bestagini et al. (2017) applied the XGBoost model to predict reservoir fluid layers based on logging data and found that the ensemble learning model shows good performance in reservoir fluid classification. Tohidi-Hosseini et al. (2016) used least-squares support vector machine (LSSVM) optimized by coupled simulated annealing (CSA) to predict the reservoir fluid and showed that the LSSVM has better learning ability after being optimized using an optimization algorithm. Sun et al. (2019) used a machine learning feed-forward neural network (FNN) and hierarchical cluster analysis (HCA) to predict reservoir fluid through natural gamma (NG) rays, high-resolution density (HRD) mapping, and single-point resistance (SPR) logging. The prediction results obtained show that their model had an R2 of 0.84, reflecting the importance of data uniformity in reservoir fluid identification. He et al. (2020) used a deep neural network (DNN) to learn logging data for reservoir fluid identification. The study adopted MAHAKIL to improve the performance of DNN on imbalanced data and obtained good prediction results (F1 of 0.601, F0.5 of 0.597, and F2 of 0.577), reflecting the need for model adjustment. Luo et al. (2022) proposed a long short-term memory network (LSTM) to characterize the time series features of logs varying with the depth domain. The kernel of the convolutional neural network (CNN) is used to slide on log curves to characterize their relationships. The innovation of Luo’s work is not only an improved AI model but also the creation of a multilevel reservoir identification process. At present, there are two challenges in reservoir fluid identification for unconventional reservoirs: 1. unconventional reservoirs show unconventional features and irregular noise information in well logging data, which require data cleaning, processing, and prevention methods for the overfitting phenomenon; 2. logging data from unconventional reservoirs usually face data imbalance problems, which makes it difficult to apply the model to fluid identification in unconventional reservoirs.
2 Methodologies
2.1 Multiverse optimizer
The multiverse optimizer (MVO) is a new method of an intelligent optimization algorithm proposed by Mirjalili et al. (2016), which has been successfully applied in various function optimizations and engineering designs (Vivek et al., 2018; Hassan and Zellagui, 2019; Jain et al., 2019; Dao et al., 2020; Zhou et al., 2022). The multiverse theory in physics is the inspiration for the MVO algorithm. In the multiverse theory, the emergence of the individual universe is the result of a single giant explosion, and multiple giant explosions have contributed to the birth of the entire multiverse population. White holes, black holes, and wormholes are the three core concepts in the multiverse theory: white holes have strong repulsion and can release all objects; black holes have extremely high gravitational forces and can absorb all objects; and wormholes connect different universes and the orbit of the transported object.
MVO is based on the principle that matter in the universe transfers from white holes to black holes through wormholes. Under the combined action of white holes, black holes, and wormholes, the entire multiverse population will eventually reach a state of convergence. MVO classifies the search process into two stages: exploration and development. The exploration of the search space is completed through the exchange of black/white holes, and the development process is completed in the way of wormholes. MVO cyclically iterates the initial universe through white hole/black hole tunnels and wormholes, in which the universe represents the feasible solution to the problem, the objects in the universe represent the components of the solution, and the expansion rate of the universe represents the fitness value of the solution. The mathematical model of the algorithm is given as follows:
Multiverse initialization: A set of random universes U is created, as shown in Equation 1.
Here,
Black and white hole mechanisms: Due to the different expansion rates of each individual universe, objects in an individual universe are transferred through white hole/black hole orbits. This process follows the roulette mechanism, as shown in Eq. 2.
where
Wormhole mechanism: Without considering the size of the expansion rate, in order to achieve local changes and improve its own expansion rate, the individual universe will stimulate the internal objects to move to the current optimal universe as shown in Eqs 3–5.
Here,
Here,
2.2 Opposition-based learning for the multiverse optimizer
In the MVO, the update of the individual mainly depends on the size of the expansion rate and is then randomly updated according to the current global optimal universe and wormhole existence probability (WEP) parameters. Since the optimal value in the early stage of the algorithm is often too far from the true value, using the global optimal universe and the updated strategy will increase the probability of the algorithm falling into the local optimal and may cause the algorithm to slow down the convergence speed. Therefore, this study introduces opposition-based learning (OBL) to improve the global search ability and algorithm stability of the MVO.
OBL was proposed by Tizhoosh (2005) to improve the convergence stability and global search ability of other algorithms by considering opposite anti-population, opposite weight, anti-behavior, anti-exploration, anti-exploitation, etc. Currently, some scholars have introduced OBL into optimization algorithms such as PSO (Wang et al., 2011), WOA (Ewees et al., 2018), SCA (Gupta and Deep, 2019), and SSA (Tubishat et al., 2020). In this study, OBL is introduced into MVO, and the OBL-MVO algorithm is established for optimizing ensemble learning. The difference between the OBL-MVO and MVO lies in three stages: negative universe initialization, anti-black hole mechanism and anti-white hole mechanism, and anti-wormhole mechanism. The OBL-MVO algorithm includes a total of six stages: initialization of the positive universe and negative universe, calculation of the expansion rate (fitness value) of the universe, black hole and white hole mechanism, anti-black hole mechanism and anti-white hole mechanism, wormhole mechanism, and anti-wormhole mechanism. The positive universe and negative universe initialization run only once, and the remaining five stages are executed in a loop. The conceptual model of the proposed algorithm is given in Figure 1.
The following section introduces the initialization of the positive universe and the negative universe, the anti-black hole mechanism, the anti-white hole mechanism, and the anti-wormhole mechanism.
Anti-multiverse establishment: In order to solve the situation of the optimal value of the MVO algorithm often being too far from the real value in the early stage, the global search ability and convergence speed of the algorithm are improved. From the establishment stage of the multiverse population, the universe is divided into a positive universe and negative universe (the number of populations is consistent with the original MVO algorithm), and the positive universe is established by Eqs 6–9:
Here,
Anti-black hole mechanism and anti-white hole mechanism: The material exchange mechanism between black holes and white holes is the exploration of the search space by MVO, but there are problems of the local search accuracy being too high and global search ability being insufficient. Therefore, the global search ability and overall stability of the algorithm are increased by setting the reverse black hole and reverse white hole mechanisms. The reverse black hole and reverse white hole mechanisms include random black hole and white hole stage and reverse object propagation stage. Both stages are controlled by a convergence factor
where
The anti-wormhole mechanism: The anti-wormhole mechanism differs from the wormhole mechanism, where the anti-wormhole mechanism improves the global search by searching backward from the universe with the lowest expansion rate. To prevent this mechanism from affecting convergence, internal objects are, therefore, only stimulated to move toward the current worst universe if it is determined that the anti-wormhole mechanism is favorable to the increase in the universe’s expansion rate.
where
2.3 Adaptive ensemble learning by opposite multiverse optimizer
The idea of ensemble learning is that even if one base learner makes a wrong prediction, other base learners can correct the error, which aims to integrate multiple base learners to improve the accuracy of prediction. The main process is to first train multiple base learners by certain rules, then combine them using an integration strategy, and finally predict the result by comprehensive judgment of all base learners. Currently, ensemble learning has been successfully applied in pattern recognition, text classification, numerical prediction (Wang et al., 2011), and other fields.
The current integration strategy can be broadly divided into two categories: one is the sequential generation of base classifiers, with strong dependencies between individual learners, represented by AdaBoost (Ying et al., 2013), and the other is the parallelized integration method, which can generate base classifiers simultaneously, without strong dependencies between individuals, represented by Bagging (Dudoit and Fridlyand, 2003).
As unconventional reservoir data are collected from different reservoirs and have different reservoir data characteristics, this study improves the accuracy of the model by accommodating as many base learners as possible into a parallelized integration method to suit the different reservoir data characteristics. However, accommodating more models can lead to overfitting (Džeroski and Ženko, 2004). Therefore, the K-fold cross-validation method and opposition-based learning are introduced to avoid overfitting.
In this study, the search for the best weights of base learners is considered an optimization problem, and the OBL-MVO algorithm is used to solve the optimization problem. The integration strategy of AIL-OMO is to divide the original dataset into several sub-datasets which are fed into each base learner in layer 1. In layer 1, each base learner outputs its own prediction results as meta features. The meta features are then used as an input to layer 2, where the OBL-MVO is used to search for the best weights of base learners.
The integration strategy used in this study is shown in Figure 2.
In order to fully exploit the individual strengths of each model, the probabilities of each reservoir predicted by each model are used as an input to train the ensemble learning model. In order to improve the accuracy and generalize the ability of the models completed by integration learning, the aforementioned OBL-MVO algorithm is introduced to reasonably optimize the integration learning process and search for the best weight of base learners.
The main contributions of this study are as follows:
(1) Collection of data and work on data collation and labeling.
(2) Proposed an algorithm OBL-MVO to optimize integration learning.
(3) Proposed an ensemble learning model AIL-OMO for the reservoir fluid classification task.
3 Experiments and results
To validate the effectiveness of our proposed method, we conducted experiments on the dataset. The following sections provide details on the model implementation, dataset processing, and final comparison results. The Flowchart of the AIL-OMO is shown in the Figure 3.
3.1 Data source and pretreatment
In this section, the data used in this paper and the preprocessing of the data are introduced. In order to build a reservoir classification model that can adapt to various unconventional features, 12,379 pieces of reservoir fluid property data from eight different wells were used for training the model. The data contain six categories of labels: dry layer, poor oil layer, water layer, oil-bearing water layer, oil–water layer, and oil layer. The data include AC (sonic interval transit time), CNL (compensated neutron logging), DEN (compensated density), GR (natural gamma ray), SP (spontaneous potential), CAL (caliper logging), and R045 (0.45 m potential resistivity). Figure 4 shows the degree of correlation between variables in reservoir fluid property data (excluding the oil field name). After obtaining the collected data, the annotation work of the dataset was carried out by analyzing the logging curves as the training data for the model.
In the data preprocessing stage, this study mainly operates on the following issues:
(1) Rejection operations for data that are missing or do not conform to common sense.
(2) For labels with labeled data, one-hot coding is used for processing. Category labels from 1 to 6 represent the dry layer, poor oil layer, water layer, oil-bearing water layer, oil–water layer, and oil layer.
(3) Data normalization of the data is used as an input to the model.
3.2 Implementation details
In this section, OBL-MVO is used to solve the optimization problem to build adaptive ensemble learning classifiers for reservoir fluid property prediction. Since the dataset of unconventional reservoirs contains various unconventional reservoir features, each base learner is differently adapted to the unconventional reservoir features. Therefore, the weight searching problem in the integration strategy is abstracted as an optimization problem to obtain and integrate a model with generalization performance on all unconventional reservoir features in the dataset. The OBL-MVO algorithm is used to optimize the weights of the base learners based on the K-fold cross-validation test set to build the AIL-OMO to avoid the overfitting of partial features for unconventional reservoirs. In this optimization problem, to avoid overfitting of the local features of unconventional reservoirs, the accuracy of the validation set in K-fold cross-validation of each base learner for each unconventional reservoir feature is obtained from the validation set in K-fold cross-validation. In this case, the weights of the base learners are encoded into the individual dimensional attributes of the population, as shown in Eq. 15. The degree of learning of each model in each unconventional reservoir feature is involved in the calculation of the adaptation value of the optimization algorithm, as shown in Eq. 16.
where
In order to show the performance of OBL-MVO on this dataset, OBL-MVO is compared with the other swarm intelligence heuristic algorithms, including whale optimization algorithm (WOA), gray wolf optimization (GWO), and MVO.
Among them, the parameter settings of each swarm intelligence heuristic algorithm are shown in Table 1.
In order to compare the four intelligent optimization algorithms, it is necessary to compare the fitness value curves of their optimization results. Since the population size of different algorithms affects the convergence of the fitness curve, in order to balance the effect of the population size of each algorithms on the fitness value, the population size for each algorithm is set to 20, 50, 100, and 200. The fitness curve results of GWO, WOA, MVO, and OBL-MVO are shown in Figure 5.
Figure 5 shows that OBL-MVO has a faster convergence rate than MVO, WOA, and GWO (the number of iterations required for the fitness function to converge and stabilize is the least); OBL-MVO shows a better convergence fitness value than MVO, WOA, and GWO, indicating that OBL-MVO has a better optimization effect than MVO, WOA, and GWO on the dataset. The corresponding weights of base learners optimized by OBL-MVO adaptive optimization are shown in Table 2.
3.3 Reservoir fluid property prediction
In this section, the predicted results of the AIL-OMO model and evaluation of its performance are presented. The AIL-OMO model was trained using the training dataset and evaluated on the test set. This paper adopts the accuracy rate as the evaluation metric, which is widely used in reservoir classification, lithology identification, and reservoir fluid identification (Moffatt and Williams, 1998; Al-Anazi and Gates, 2010; Boyd et al., 2013; Onwuchekwa, 2018).
Figure 6 depicts the results of using seven machine learning algorithms to predict the reservoir fluid, including the AIL-OMO model proposed in this paper. The accuracy of the AIL-OMO model surpasses that of the other six models, achieving an accuracy rate of 92.75%, followed by the SVC and linear SVC models in well 1. Figure 7 presents the reservoir fluid prediction results of the seven machine learning models in well 2. The accuracy rates from high to low are given as follows: AIL-OMO (81.93%), logistic regression (70.37%), linear SVC (69.30%), linear discriminant analysis (69.26%), KNN (69.11%), AdaBoost (66.45%), and SVC (64.37%). Figure 8 shows the reservoir fluid prediction results of the seven machine learning models in well 3. Among the prediction results of each model, the accuracy rate of the AIL-OMO model is better than that of the other models, with an accuracy rate of 88.72%, followed by SVC with an accuracy rate of 83.03%. The worst model is AdaBoost, with an accuracy rate of only 65.57%. From the experimental results, it can be interpreted that integrating multiple weak learners into one strong learner using the OBL-MVO algorithm leads to a significant improvement in the prediction accuracy of the original model. Therefore, the proposed AIL-OMO model has high accuracy and is more conducive to the prediction of reservoir fluid.
Figure 9 shows the change in the classification accuracy of the AIL-OMO model with different party sizes.
FIGURE 9. Reservoir fluid property prediction by AIL-OMO (classification label: dry layer (1), poor oil layer (2), water layer (3), oil-bearing water layer (4), oil–water layer (5), and oil layer (6)).
Figure 9 shows the classification accuracy of different reservoir fluids by AIL-OMO. Each value in the figure represents the ratio of the number of reservoir fluids identified by AIL-OMO to the actual number of reservoir fluids in the dataset. Figure 6 shows that the results of AIL-OMO exhibit slight fluctuations under different party sizes. The classification accuracy for the poor oil layer, water layer, and oil-bearing water layer is relatively low. This can be attributed to the smaller percentage of poor reservoirs, water layers, and oil-bearing water layers in the dataset.
Table 3 presents the comparison results of the accuracy of AIL-OMO and the other six machine learning models. Among them, AIL-OMO with a party size of 100 achieves the highest accuracy of 85%. This model demonstrates the fewest classification errors for each reservoir. However, when the party size is increased to 200, the accuracy of AIL-OMO decreases to 78%. This decrease in accuracy can be attributed to the party size being too large, causing overfitting of the model to the data.
4 Discussion
Considering the characteristics of ensemble learning, the MVO algorithm may promote the overfitting learning phenomenon in the fluid category with small sample data. Therefore, the more models the AIL-OMO integrates, the more the corresponding training data samples should be, so that MVO can achieve the optimization effect of ensemble learning.
For each model, the identification of oil-bearing water layer and water layer are not accurate enough, which may be due to the classification confusion caused by noise interference in the logging curves of the water layer in unconventional reservoirs. In the future, we aim to decrease the number of base models to improve the operational speed and enhance our model’s performance by collecting data from various reservoir types for testing.
5 Conclusion
This study focuses on logging data for unconventional reservoirs, which exhibit unconventional characteristics and data imbalance. To address these challenges, this paper proposes a new model that combines 12 base learners using the multiverse optimizer. The AIL-OMO model combines multiple weak learners into one strong learner. The advantage of our ensemble learning lies in its ability to ensure the diversity of weak classifiers, resulting in better prediction results than those obtained by other single learner models. Compared to other widely used models, the new ensemble learning model proposed in this paper achieves high accuracy on classifying dry layers, oil–water layers, and oil layers, with accuracy rates of 94.33%, 90.46%, and 90.66%, respectively. This indicates that the proposed model exhibits the highest accuracy and better generalization for fluid identification in unconventional oil reservoirs.
Data availability statement
The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.
Author contributions
YZ: writing—original draft and resources; XZ: methodology and supervision; YS: visualization and investigation; AG: data curation and conceptualization; ML: validation. All authors contributed to the article and approved the submitted version.
Funding
This work was financially supported by the Major Scientific and Technological Projects of CNPC under Grant ZD2019-183-004.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Al-Anazi, A., and Gates, I. D. (2010). A support vector machine algorithm to classify lithofacies and model permeability in heterogeneous reservoirs. Eng. Geol. 114, 267–277. doi:10.1016/j.enggeo.2010.05.005
Baarimah, S. O., Al-Aidroos, N. M., and Ba-Jaalah, K. S. (2019). Using chemical composition of crude oil and artificial intelligence techniques to predict the reservoir fluid properties. Int. Conf. Intell. Comput. Eng. (ICOICE), 1–5. doi:10.1109/ICOICE48418.2019.9035135
Bestagini, P., Lipari, V., and Tubaro, S. (2017). A machine learning approach to facies classification using well logs Seg technical program expanded abstracts 2017. Houston, Texas, United States: Society of Exploration Geophysicists, 2137–2142.
Boyd, K., Eng, K. H., and Page, C. D. “Area under the precision-recall curve: Point estimates and confidence intervals,” in Proceedings of the Joint European conference on machine learning and knowledge discovery in databases, Prague, Czech Republic, September 2013 (Springer), 451–466.
Dao, T., Yu, J., Nguyen, T., and Ngo, T. (2020). A hybrid improved MVO and FNN for identifying collected data failure in cluster heads in WSN. IEEE Access 8, 124311–124322. doi:10.1109/access.2020.3005247
Das, B., and Chatterjee, R. (2018). Well log data analysis for lithology and fluid identification in Krishna-Godavari Basin, India. Arabian J. Geosciences 11, 231–312. doi:10.1007/s12517-018-3587-2
Dudoit, S., and Fridlyand, J. (2003). Bagging to improve the accuracy of a clustering procedure. Bioinformatics 19, 1090–1099. doi:10.1093/bioinformatics/btg038
Džeroski, S., and Ženko, B. (2004). Is combining classifiers with stacking better than selecting the best one? Mach. Learn 54, 255–273. doi:10.1023/b:mach.0000015881.36452.6e
Ewees, A. A., Abd Elaziz, M., and Houssein, E. H. (2018). Improved grasshopper optimization algorithm using opposition-based learning. Expert Syst. Appl. 112, 156–172. doi:10.1016/j.eswa.2018.06.023
Guerreiro, J. N. C., Barbosa, H. J. C., Garcia, E. L. M., Loula, A. F. D., and Malta, S. M. C. (1998). Identification of reservoir heterogeneities using tracer breakthrough profiles and genetic algorithms. SPE Reserv. Eval. Eng. 1, 218–223. doi:10.2118/39066-pa
Gupta, S., and Deep, K. (2019). A hybrid self-adaptive sine cosine algorithm with opposition based learning. Expert Syst. Appl. 119, 210–230. doi:10.1016/j.eswa.2018.10.050
Hassan, H. A., and Zellagui, M. “MVO algorithm for optimal simultaneous integration of DG and DSTATCOM in standard radial distribution systems based on technical-economic indices,” in Proceedings of the 2019 21st International Middle East Power Systems Conference (MEPCON), Cairo, Egypt, December 2019, 277–282.
He, M., Gu, H., and Wan, H. (2020). Log interpretation for lithology and fluid identification using deep neural network combined with MAHAKIL in a tight sandstone reservoir. J. Pet. Sci. Eng. 194, 107498. doi:10.1016/j.petrol.2020.107498
Jain, G., Yadav, G., Prakash, D., Shukla, A., and Tiwari, R. (2019). MVO-based path planning scheme with coordination of UAVs in 3-D environment. J. Comput. Sci. 37, 101016. doi:10.1016/j.jocs.2019.07.003
Luo, G., Xiao, L., Liao, G., Luo, S., Shao, R., Zhou, J., et al. (2022). Multi-level reservoir identification with logs based on machine learning SPWLA 63rd annual logging symposium. Richardson, TX, United States: OnePetro.
Mirjalili, S., Mirjalili, S. M., and Hatamlou, A. (2016). Multi-verse optimizer: A nature-inspired algorithm for global optimization. Neural Comput. Appl. 27, 495–513. doi:10.1007/s00521-015-1870-7
Moffatt, B. J., and Williams, J. M. (1998). Identifying and meeting the key needs for reservoir fluid properties A multi-disciplinary approach spe annual technical conference and exhibition. Richardson, TX, United States: OnePetro.
Onwuchekwa, C. (2018). Application of machine learning ideas to reservoir fluid properties estimation SPE Nigeria Annual International Conference and Exhibition. Richardson, TX, United States: OnePetro.
Sun, J., Li, Q., Chen, M., Ren, L., Huang, G., Li, C., et al. (2019). Optimization of models for a rapid identification of lithology while drilling - a win-win strategy based on machine learning. J. Pet. Sci. Eng. 176, 321–341. doi:10.1016/j.petrol.2019.01.006
Tizhoosh, H. R. “Opposition-based learning: A new scheme for machine intelligence,” in Proceedings of the International conference on computational intelligence for modelling, control and automation and international conference on intelligent agents, web technologies and internet commerce (CIMCA-IAWTIC'06), Vienna, Austria, November, 2005, 695–701.
Tohidi-Hosseini, S., Hajirezaie, S., Hashemi-Doulatabadi, M., Hemmati-Sarapardeh, A., and Mohammadi, A. H. (2016). Toward prediction of petroleum reservoir fluids properties: A rigorous model for estimation of solution gas-oil ratio. J. Nat. Gas. Sci. Eng. 29, 506–516. doi:10.1016/j.jngse.2016.01.010
Tubishat, M., Idris, N., Shuib, L., Abushariah, M. A., and Mirjalili, S. (2020). Improved Salp Swarm Algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst. Appl. 145, 113122. doi:10.1016/j.eswa.2019.113122
Vivek, K., Deepak, M., Mohit, J., Asha, R., and Vijander, S. (2018). Development of multi-verse optimizer (mvo) for labview intelligent communication, control and devices. Berlin, Germany: Springer, 731–739.
Wang, H., Wu, Z., Rahnamayan, S., Liu, Y., and Ventresca, M. (2011). Enhancing particle swarm optimization using generalized opposition-based learning. Inf. Sci. 181, 4699–4714. doi:10.1016/j.ins.2011.03.016
Wang, P., Chen, X., Wang, B., Li, J., and Dai, H. (2020). An improved method for lithology identification based on a hidden Markov model and random forests. Geophysics 85, IM27–IM36. doi:10.1190/geo2020-0108.1
Yang, H., Pan, H., Ma, H., Konaté, A. A., Yao, J., and Guo, B. (2016). Performance of the synergetic wavelet transform and modified K-means clustering in lithology classification using nuclear log. J. Pet. Sci. Eng. 144, 1–9. doi:10.1016/j.petrol.2016.02.031
Ying, C., Qi-Guang, M., Jia-Chen, L., and Lin, G. (2013). Advance and prospects of AdaBoost algorithm. Acta Autom. Sin. 39, 745–758. doi:10.1016/s1874-1029(13)60052-x
Zhang, T., and Sun, S. (2021). Thermodynamics-informed neural network (TINN) for phase equilibrium calculations considering capillary pressure. Energies 14, 7724. doi:10.3390/en14227724
Keywords: reservoir identification, ensemble learning, reverse learning, machine learning, unconventional reservoirs
Citation: Zhang Y, Zhang X, Sun Y, Gong A and Li M (2023) An adaptive ensemble learning by opposite multiverse optimizer and its application in fluid identification for unconventional oil reservoirs. Front. Earth Sci. 11:1116664. doi: 10.3389/feart.2023.1116664
Received: 10 January 2023; Accepted: 12 July 2023;
Published: 26 July 2023.
Edited by:
Mehdi Ostadhassan, Northeast Petroleum University, ChinaReviewed by:
Tao Zhang, King Abdullah University of Science and Technology, Saudi ArabiaSaid Gaci, Algerian Petroleum Institute, Algeria
Copyright © 2023 Zhang, Zhang, Sun, Gong and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: An Gong, 414625329@qq.com