- 1Key Laboratory of Deep Oil and Gas, China University of Petroleum (East China), Qingdao, China
- 2Geophysical Research Institute, School of Earth and Space Sciences, University of Science and Technology of China, Hefei, China
With high imaging accuracy, high signal-to-noise ratio, and good amplitude balance, least-squares reverse time migration (LSRTM) is an imaging algorithm suitable for deep high-precision oil and gas exploration. However, the computational costs limit its large-scale industrial application. The difference between traditional reverse time migration (RTM) and LSRTM is whether to eliminate the effect of the Hessian operator or not while solving Hessian matrix explicitly or eliminating the effect of the Hessian matrix implicitly has a very high requirement on computation or storage capacity. We simulate the inverse Hessian by training a cycle-consistent generative adversarial network (cycleGAN) to construct a mapping relationship between the RTM results and the true reflectivity models. The trained network is directly applied to the RTM imaging results, which improves the imaging quality while significantly reducing the calculation time. We select three velocity models and two velocity models respectively to generate the training and validation data sets, where the validation data is not involved in the training process. The prediction results on the validation data sets show that the trained network significantly improves the imaging quality with almost no additional in computational effort. Finally, we apply the network trained with only synthetics to the field data. The predicted results confirm the effectiveness and good generalization of the proposed method.
Introduction
For the high-precision imaging problem of complex geological formations and deep reservoirs, the reverse time migration (RTM) based on the two-way wave-equation is considered to be a better imaging method (Baysal et al., 1983; Chang, 1987). Essentially, RTM is only the conjugate transpose of the forward operator, not its inverse (Lailly, 1983). Therefore, influenced by the limited acquisition aperture, complex subsurface structure, and limited seismic wave bandwidth, the RTM results suffer from migration noise, non-uniform illumination, and insufficient resolution (Cheng et al., 2021), which prevents it from imaging deep and unconventional oil and gas reservoirs finely. In addition, conventional RTM can only image the subsurface discontinuities while the accurate amplitude information of the real formation lithology changes can not be guaranteed (Chen and Zhang, 2012).
To eliminate the effect of the Hessian matrix, the imaging problem of RTM can be solved as a least-squares inverse problem to suppress the migration noise, increase the resolution, and improve the amplitude fidelity, which is the least-squares reverse time migration (Tarantola, 1987; Chavent and Plessix, 1999; Nemeth et al., 1999; Valenciano et al., 2006). The least-squares reverse time migration (LSRTM) can be solved in model space (Hu et al., 2001; Rickett, 2003; Guitton, 2004) or in data space (Tarantola, 1987; Schuster, 1993; Chavent and Plessix, 1999; Nemeth et al., 1999). The LSRTM in model space requires solving the Hessian matrix explicitly and using its inverse to deblurring the RTM imaging results. Directly solving the full Hessian matrix in practical applications is extremely demanding in terms of computation and storage (Tang, 2009). Therefore, some researchers have proposed that the Hessian matrix can be approximated as a diagonal matrix under the assumption of high-frequency asymptotics and infinite aperture (Beylkin, 1985; Chavent and Plessix, 1999; Plessix and Mulder, 2004; Symes, 2008). However, this approximation is not suitable for the limited frequency range and acquisition geometry, and the Hessian matrix in this case is no longer diagonally dominant (Pratt et al., 1998; Chavent and Plessix, 1999; Plessix and Mulder, 2004; Valenciano et al., 2006). In regions below the high-velocity body or salt dome where illumination is lacking, the energy on the diagonal of the Hessian matrix is scattered along the non-diagonal direction (Albertin et al., 2004; Valenciano, 2008). Therefore, the diagonal Hessian matrix is of limited use for deblurring the migration imaging results, especially in regions with severe energy attenuation of seismic waves. Some non-diagonal elements of the Hessian matrix can be properly computed to compensate for the illumination unevenness in regions with severe absorption attenuation and enhance the deblurring effect (Albertin et al., 2004; Valenciano, 2008). However, solving the non-diagonal elements of the matrix will significantly increase the computational effort.
The LSRTM in data space is used to find the reflectivity model that best matches the observed seismic data by iterative algorithms such as the steepest descent method and the conjugate gradient method (Nemeth et al., 1999). Instead of solving the huge Hessian matrix directly, this approach implicitly removes the effect of the Hessian matrix by iteratively updating the reflectivity model. However, iterative solving causes a large amount of computation, and the convergence rate depends on suitable preconditioning constraints (Tang, 2009; Wu et al., 2021; Wu et al., 2022; Yao et al., 2022). To improve the computational efficiency of LSRTM, one approach is to compress the massive shot data into several super gathers to reduce the computational effort, which is the multisource migration method (Romero et al., 2000). To suppress the multi-source crosstalks, some researchers proposed the shot encoding techniques, such as amplitude encoding (Hu et al., 2016), phase encoding (Schuster et al., 2011; Li et al., 2014; Li et al., 2017) and plane wave encoding (Zhang et al., 2005; Dai and Schuster, 2013; Li et al., 2018). These schemes reduce the correlation of crosstalk noise in the imaging results of different super gathers by constructing suitable encoding functions and finally stack the imaging results to suppress the noise (Etgen, 2005; Zhang et al., 2005). However, when the observation system is not fixed, the source wavefield of the multisource forward simulation and the super gather synthesized from the observation data cannot be matched, resulting in slow convergence (Li et al., 2018). Liu and Peter (2018) propose a one-step LSRTM method to design a deblurring preconditioner using the Wiener filter, reducing the computational effort to twice the RTM. In summary, LSRTM in model space requires certain calculations and significant storage costs to solve the Hessian matrix explicitly, while LSRTM in data space requires a large amount of computation to solve the reflectivity iteratively and does not require much storage.
In recent years, with the popularity of artificial intelligence, many researchers introduced deep learning techniques to solve geophysical problems such as fault segmentation (Wu et al., 2019), seismic data denoising (Yu et al., 2019; Saad and Chen, 2020), and first-arrival picking (Hu et al., 2019; Duan and Zhang, 2020). These solutions use techniques such as DNN, CNN, RNN, etc. These networks can be broadly classified into two categories based on the application scenarios: discriminative models that solve classification, identification, or localization problems, and generative models that solve the mapping relationship between data from two different domains. The generative adversarial network (GAN) combines the advantages of both types of networks by using a generative model to produce data from another domain and then using a discriminative model to evaluate the generated data (Goodfellow et al., 2014). Through adversarial training, GAN achieves state-of-the-art in solving the problem of style transfer between data from different domains. The cycleGAN (Zhu et al., 2017) consists of two GANs that ensure the structural consistency of the input and output data through cyclic consistency, taking full advantage of the mapping relationship between the source and target domain data. This network structure is well suitable for transferring the RTM data into a true reflectivity model and eliminating the blurring effect of Hessian matrix.
In this study, we propose to approximate the inverse Hessian by using cycleGAN for the deblurring purpose. The cycleGAN is trained by generating data sets of RTM imaging results with reflectivity models. The trained network is directly applied to the migration images to output high-quality imaging profiles. The training stage of the network consumes a certain computational effort and has an almost negligible computational cost. The proposed method is validated on both synthetic data and field data.
Methodology
Migration theory
For a given reflectivity model
The goal of seismic imaging is to find the exact
where
The solution of the true reflectivity model
Applying the transpose forward operator
Therefore, the precise least-squares solution of the reflectivity model
where
Analysis of Hessian matrix
The expression of the conventional two-dimensional acoustic wave equation is:
where
where
According to Eq. 6, it can be seen that the RTM results are missing the deblurring filtering of the inverse Hessian matrix with respect to the real reflectivity models. And according to Eq. 8, it can be concluded that the wavelet term and the Green’s function term in the expression of the Hessian matrix affect the imaging resolution and wavefield propagation, respectively. Therefore, the migrated images of RTM exhibit low imaging resolution, uneven illumination, and migration noise due to the lack of the inverse Hessian filtering. By solving the inverse Hessian matrix and applying it to the imaging results of the RTM, the true reflectivity model can be obtained, which is the least-squares reverse time migration in the model space. Theoretically, the LSRTM in model space only requires a certain computational cost to compute the inverse of the Hessian matrix in advance, and then the true reflectivity model can be acquired quickly. However, it is shown that solving the Hessian matrix requires storing the Green’s function, which is very large in practical applications, especially for 3D seismic data. Another approach is the LSRTM in the data space, which is an iterative solution to continuously update the reflectivity model so that the error between the simulated and observed data reaches the desired value. The LSRTM in the data domain does not take into account the storage cost, but it is computationally intensive, about 2N times more than the conventional migration (N denotes the number of iterations). Moreover, the convergence of the LSRTM in data space is very slow without proper gradient preconditioning.
Our goal is to approximate the inverse of the Hessian matrix at a small cost. In the proposed method, we train cycleGAN to simulate the inverse Hessian and apply the trained network to process the RTM imaging data and play the role of the deblurring operator. First, we select some velocity models and generate the true reflectivity model
Network and loss function
The architecture of cycleGAN
GAN was proposed by Goodfellow et al. (2014), and it has good applicability to the problem of the source-to-target domain transformation. Conventional neural networks are used to find an optimal mapping relationship from input to output by backpropagation algorithms. The nature of the network can be viewed as a complex higher-order polynomial. GAN generally consists of two parts of the network, one as a generator and the other as a discriminator. During the training process of the network, the generator acts as a regular neural network, converting the input into data similar to the label, while the discriminator takes the output data of the generator as input and tries to distinguish the generated data from the real data in the training sets. Through this adversarial training, both the generator and the discriminator are continuously optimized, and eventually, we get the desired generator as a prediction network.
CycleGAN consists of 2 mirror-symmetric networks
FIGURE 2. The specific structure of the generator, consists of 2 convolutional layers forming the encoder, 7 Resnet blocks as the converter, and 2 deconvolutional layers as the decoder.
FIGURE 3. The detailed architecture of the discriminator, 4 convolutional layers are used to extract the features of the input data, and finally, a convolutional layer is added to output the discriminant score.
Loss function analysis
The loss function is a measure of how good the output of the network is, and it is also the basis for correcting the tensor gradient in the backpropagation process. For a given size training batch, assuming the data distribution of the source domain is
During training, the generator
By the above constraint of the objective function, GAN can theoretically map any data from the source domain to the target domain. However, only this constraint may result in the generator just generating a result with reflectivity characteristics, while the structural content is not the same as the source domain data. That is, only style transfer is achieved by training, and there is no guarantee that the content does not change. For any source domain data
With the above loss function constraint, any data in the source or target domain can be effectively restored to the original data features in the end after two generator cycles. However, when the network is trained for application, the situation that the style completes the migration but the content changes still occurs. The analysis shows that only one generator is used as the prediction network when applied, and the cyclic consistency error function does not impose constraints on the intermediate variables. Therefore, we introduce the identity loss function to constrain the intermediate variables, and its expression is (Taigman et al., 2016; Benaim and Wolf, 2017; Zhu et al., 2017):
Therefore, the full loss function when training the cycleGAN is shown as follows:
Data sets preparation and training
Training data sets production
As shown in Figures 4A–C, the Sigsbee2A model, SEAM model, and Pluto model are selected as the base models to produce the training data sets. The migrated images corresponding to the three models are generated by 2-8 finite-difference forward simulation and reverse time migration as shown in Figures 4D–F. Meanwhile, we produce the reflectivity model based on the velocity model as shown in Figures 4H–J. Considering the large band range variability between the reflectivity model and the migration imaging results, we adopt the convolution approach and conduct it with the reflectivity and the Ricker wavelet as the target domain data. Finally, we apply data enhancement such as cropping and resampling to generate 300 pairs of training data. Some of the training data are shown in Figure 5. The first row shows the RTM imaging data in the source domain, and the second row shows the convolution results in the target domain.
FIGURE 4. Three velocity models for producing training data sets in the source and target domains. (A) Sigsbee2A velocity model, (B) SEAM velocity model, and (C) Pluto velocity model. The corresponding RTM imaging results (D–F) and reflectivity models (H–J) for making the input and output samples respectively.
FIGURE 5. Four pairs of representative samples from the training data sets. The first row (A–D) are the migrated images in the source domain, and the second row (E–H) denotes the reflectivity in the target domain.
Training process
For
The number of epochs of the network training is set to 200 with a learning rate of 0.0002, which gradually decreased after the 100th epoch. The network is weighted with a cyclic consistency loss of 10 and the weight of the identity loss is set to 5. The training is based on the PyTorch platform, using a single NVIDIA RTX5000 GPU, and the total training time is 10 h.
Numerical tests on validation sets
BP 2.5D model
Figure 6A shows the BP2.5D velocity model with the reflectivity model shown in Figure 6B. We perform a finite-difference forward simulation of the velocity model in Figure 6A, then smooth the velocity model into migration velocity and perform reverse time migration to generate the imaging profile in Figure 6C. There are obvious low-frequency noise and acquisition footprints in the migrated image of the RTM, and the imaging energy is discontinuous in the region of violent tectonic undulations, meanwhile, the illumination is uneven on both sides of the model. We input the imaging result of Figure 6C into the trained network, and the obtained prediction result is shown in Figure 6D. The imaging quality in the predicted result is significantly improved, the surface acquisition footprints are eliminated very cleanly, and the low-frequency noise is well removed. The trained network compensates well for the uneven illumination in the migration result, and the overall imaging energy distribution is relatively uniform. It can be noted that the network is not completely accurate in predicting the result at the depth of 2 KM on the left side of the model since the imaging energy of the original RTM profile is very heterogeneous, resulting in the network not being able to predict it well. Figure 7 shows an enlarged view of the red dashed boxed area in Figure 6, with the first column showing the RTM imaging results, the second column denoting the network prediction results, and the third column representing the true reflectivity models. Comparing the first set of results (Figures 7A–C), it can be seen that the proposed method can better play the role of the deblurring operator and significantly improve the imaging quality. The second set of results (Figures 7D–F) shows that the trained network can not fully reduce the image of the RTM to the true reflectivity model, and it relies to some extent on the structure information of the original migration image, which differs from the real reflectivity model in some details, but generally maintains the consistency with the real reflectivity model.
FIGURE 6. BP 2.5D velocity model (A) and the corresponding reflectivity model (B). Input the RTM imaging result (C) into the well-trained network, the prediction result is shown in (D).
FIGURE 7. The zoom views of the red dashed boxed area in Figure 6. The first column represents the RTM results (A,D), the second column denotes the predictions (B,E) of the network and the third column is the true reflectivity model (C,F).
Marmousi model
Figure 8A shows the Marmousi velocity model, and Figure 8B shows its corresponding reflectivity model. The imaging result obtained after finite-difference simulation and reverse time migration of Figure 8A is shown in Figure 8C. The migrated image of RTM has more low-frequency noise in the shallow layer due to the blurring filtering of Hessian matrix. At the same time, the imaging resolution of the model bottom structure is low due to the influence of the deep high-speed layer shielding. Figure 8D shows the result of network prediction. Compared with Figure 8C, it can be concluded that the global resolution of the predicted imaging result is improved, and the reflection layer is finer. Most of the low-frequency noise in the shallow layer is removed, with a small amount of residual on both sides, and the trained network is also very clean for removing the acquisition footprints in the profile. The imaging quality of the deeper layers is also improved, however, some artifacts appear since some stray imaging energies in the result of the RTM are mistaken by the network as layers and recovered. Figure 9 shows an enlarged view of the red dashed boxed area in Figure 8. Figure 9A shows a shallow zoom of the RTM imaging result, where strong low-frequency noise and acquisition footprints exist within 0.5 KM of depth. These clutter energies are barely visible in the predicted result in Figure 9B, and the imaging energies at the interface are more converged. Comparing the second set of zooming results, it can be seen that the resolution of the result predicted by the network has improved significantly and the overall imaging energy is more balanced.
FIGURE 8. The Marmousi velocity model (A) and the corresponding reflectivity model (B). Input the RTM imaging result (C) into the well-trained network, the prediction result is shown in (D).
FIGURE 9. The zoom views of the red dashed boxed area in Figure 8. The first column represents the RTM results (A,D), the second column denotes the predictions (B,E) of the network and the third column is the true reflectivity model (C,F).
Field data test
We test the effectiveness of the proposed method with field data. We take the velocity model in Figure 10A as the migration velocity and perform the reverse time migration to obtain the imaging result shown in Figure 10B. We approximate the deblurring operator with the network previously trained on synthetic data, which acts directly on the migrated image of the field data. We input the migration result in Figure 10B into the network, and the prediction result output by the generator
FIGURE 10. (A) migration velocity model, (B) RTM imaging result, and (C) the prediction result of the network trained only by synthetic data sets.
Conclusion
We propose a fast least-squares reverse time migration method based on cycleGAN. In the proposed method, cycleGAN is used to approximate the inverse Hessian matrix to play the role of a deblurring operator that acts directly on the RTM imaging results. By constructing a training data sets consisting of source domain data (migration imaging results) and target domain data (true reflectivity model), two GANs are trained separately from the two domains, and a generator mapping the source domain to the target domain is taken as the final prediction network after adversarial training. Synthetic data tests demonstrate the effectiveness of the proposed method in removing low-frequency noise, improving resolution, balancing illumination, and compensating for deep imaging energy. Applying the network trained only with synthetics directly to the field data, the prediction result shows that the proposed method has good generalization and good applicability to data sets with large variability. One point worth noting is that in the prediction result of the field data, the recovery of high steepness structures by the method is not very well. This problem can be solved by optimizing the label-making scheme and reducing the spatial sampling interval. Compared with the massive storage requirement of the Hessian matrix by the LSRTM in the model space, the storage demands of the proposed method are not high. Compared with the LSRTM in data space, the proposed method only consumes certain computational resources in the training stage and requires almost no computational costs in the prediction process.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author Contributions
YH and JH contributed to the idea and methodology, YH was responsible for the training data production, model and field data testing, and the writing of this manuscript, and YM checked and polished the manuscript. All authors have read the manuscript and agreed to publish it.
Funding
This research is supported by the National Key R&D Program of China (no. 2019YFC0605503), the National Outstanding Youth Science Foundation (no. 41922028), and the Major Scientific and Technological Projects of CNPC (no. ZD 2019-183-003).
Acknowledgments
The authors are grateful to the editors and reviewers for their review of this manuscript. Meanwhile, the authors would like to thank the PyTorch platform for providing excellent deep learning tools.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Albertin, U., Yingst, D., Kitchenside, P., and Tcheverda, V. (2004). Expanded Abstracts: SEG, 949–952.True-amplitude beam migration: 74st annual international meeting
Baysal, E., Kosloff, D. D., and Sherwood, J. W. C. (1983). Reverse time migration. Geophysics 48 (11), 1514–1524. doi:10.1190/1.1441434
Benaim, S., and Wolf, L. (2017). One-sided unsupervised domain mapping. Adv. Neural Inf. Process. Syst., 752–762.
Beylkin, G. (1985). Imaging of discontinuities in the inverse scattering problem by inversion of a causal generalized Radon transform. J. Math. Phys. 26 (1), 99–108. doi:10.1063/1.526755
Chang, W. F., and McMechan, G. A. (1987). Elastic reverse‐time migration. Geophysics 52 (3), 1365–1375. doi:10.1190/1.1442249
Chavent, G., and Plessix, R. E. (1999). An optimal true‐amplitude least‐squares prestack depth‐migration operator. Geophysics 64 (2), 508–515. doi:10.1190/1.1444557
Chen, S. C., and Zhang, B. (2012). Steeply-dipping structures migration based on the wavefield vertical- and horizontal-extrapolation. Chinese J. Geophys. 55 (4), 1300–1306.
Cheng, J. B., Wang, H. Z., and Ma, Z. T. (2021). Pre-stack depth migration with finite-difference method in frequency-space domain. Chinese J. Geophys. 44 (3), 389–395.
Dai, W., and Schuster, G. T. (2013). Plane-wave least-squares reverse-time migration. Geophysics 78 (4), S165–S177. doi:10.1190/geo2012-0377.1
Duan, X. D., and Zhang, J. (2020). Multitrace first-break picking using an integrated seismic and machine learning method. Geophysics 85 (4), WA269–WA277. doi:10.1190/geo2019-0422.1
Etgen, J. T. (2005). Expanded Abstracts: SEG.How many angles do we really need for delayed-shot migration? 75st annual international meeting
Gerhard Pratt, R. G., Shin, C., and Hicks, G. J. (1998). Gauss-Newton and full Newton methods in frequency-space seismic waveform inversion. Geophys. J. Int. 133 (2), 341–362. doi:10.1046/j.1365-246x.1998.00498.x
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., et al. (2014). Generative adversarial nets. Adv. Neural Inf. Process. Syst., 2672–2680.
Guitton, A. (2004). Amplitude and kinematic corrections of migrated images for nonunitary imaging operators. Geophysics 69 (4), 1017–1024. doi:10.1190/1.1778244
Hu, J. G., Wang, H. Z., Fang, Z. Y., Li, T. C., and Zhang, J. N. (2016). Efficient amplitude encoding least-squares reverse time migration using cosine basis. Geophys. Prospect. 64 (6), 1483–1497. doi:10.1111/1365-2478.12356
Hu, J., Schuster, G. T., and Valasek, P. A. (2001). Poststack migration deconvolution. Geophysics 66 (3), 939–952. doi:10.1190/1.1444984
Hu, L. L., Zheng, X. D., Duan, Y. T., Yan, X. F., Hu, Y., Zhang, X. L., et al. (2019). First-arrival picking with a U-net convolutional network. Geophysics 84 (6), U45–U57. doi:10.1190/geo2018-0688.1
Lailly, P. (1983). The seismic inverse problem as a sequence of before stack migrations. Conf. Inverse Scatt. Theory Appl. 15 (2), 206–220.
Li, C., Huang, J. P., Li, Z. C., and Wang, R. R. (2018). Plane-wave least-squares reverse time migration with a preconditioned stochastic conjugate gradient method. Geophysics 83 (1), S33–S46. doi:10.1190/geo2017-0339.1
Li, C., Huang, J. P., Li, Z. C., and Wang, R. R. (2017). Preconditioned prestack plane-wave least squares reverse time migration with singular spectrum constraint. Appl. Geophys. 14 (1), 73–86. doi:10.1007/s11770-017-0599-8
Li, C., Huang, J. P., Li, Z. C., Wang, R. R., and Sun, M. M. (2018). Least-squares reverse time migration with a hybrid stochastic conjugate gradient method. Oil Geophys. Prospect. 53 (6), 1188–1197.
Li, Z. C., Guo, Z. B., and Tian, K. (2014). Least-squares reverse time migration in visco-scoustic media. Chin. J. Geophys. 57 (1), 214–228.
Liu, Q. C., and Peter, D. (2018). One-step data-domain least-squares reverse time migration. Geophysics 83 (4), R361–R368. doi:10.1190/geo2017-0622.1
Nemeth, T., Wu, C., and Schuster, G. T. (1999). Least‐squares migration of incomplete reflection data. Geophysics 64 (1), 208–221. doi:10.1190/1.1444517
Plessix, R. E., and Mulder, W. A. (2004). Frequency-domain finite-difference amplitude-preserving migration. Geophys. J. Int. 157 (3), 975–987. doi:10.1111/j.1365-246x.2004.02282.x
Rickett, J. E. (2003). Illumination‐based normalization for wave‐equation depth migration. Geophysics 68 (4), 1371–1379. doi:10.1190/1.1598130
Romero, L. A., Ghiglia, D. C., Ober, C. C., and Morton, S. A. (2000). Phase encoding of shot records in prestack migration. Geophysics 65 (2), 426–436. doi:10.1190/1.1444737
Saad, O. M., and Chen, Y. K. (2020). Deep denoising autoencoder for seismic random noise attenuation. Geophysics 85 (4), V367–V376. doi:10.1190/geo2019-0468.1
Schuster, G. T. (1993). Expanded Abstracts: SEG, 110–113.Least-squares cross-well migration: 63st annual international meeting
Schuster, G. T., Wang, X., Huang, Y., Dai, W., and Boonyasiriwat, C. (2011). Theory of multisource crosstalk reduction by phase-encoded statics. Geophys. J. Int. 184 (3), 1289–1303. doi:10.1111/j.1365-246x.2010.04906.x
Symes, W. W. (2008). Approximate linearized inversion by optimal scaling of prestack depth migration. Geophysics 73 (2), R23–R35. doi:10.1190/1.2836323
Taigman, Y., Polyak, A., and Wolf, L. (2016). Unsupervised cross-domain image generation: arXiv preprint arXiv:1611.02200.
Tang, Y. (2009). Target-oriented wave-equation least-squares migration/inversion with phase-encoded Hessian. Geophysics 74 (6), WCA95–WCA107. doi:10.1190/1.3204768
Tarantola, A. (1987). Inverse problem theory. Methods for data fitting and model parameter estimation. Elsevier.
Valenciano, A. A., Biondi, B., and Guitton, A. (2006). Target-oriented wave-equation inversion. Geophysics 71 (4), A35–A38. doi:10.1190/1.2213359
Wu, B., Yao, G., Cao, J. J., Wu, D., Li, X., Liu, N. C., et al. (2022). Huber inversion-based reverse-time migration with de-primary imaging condition and curvelet-domain sparse constraint. Petroleum Sci. 18. doi:10.1016/j.petsci.2022.03.004
Wu, D., Wang, Y. H., Cao, J. J., da Silva, N. V., and Yao, G. (2021). Least-squares reverse-time migration with sparsity constraints. J. Geophys. Eng. 18, 304–316. doi:10.1093/jge/gxab015
Wu, X. M., Liang, L. M., Shi, Y. Z., and Fomel, S. (2019). FaultSeg3D: Using synthetic data sets to train an end-to-end convolutional neural network for 3D seismic fault segmentation. Geophysics 84 (3), IM35–IM45. doi:10.1190/geo2018-0646.1
Yao, G., Wu, B., da Silva, N. V., Debens, H. A., Wu, D., Cao, J. J., et al. (2022). Least-squares reverse time migration with a multiplicative Cauchy constraint. Geophysics 87 (3), S151–S167. doi:10.1190/geo2021-0183.1
Yu, S. W., Ma, J. W., and Wang, W. L. (2019). Deep learning for denoising. Geophysics 84 (6), V333–V350. doi:10.1190/geo2018-0668.1
Zhang, Y., Sun, J., Notfors, C., Gray, S., Chernis, L., Young, J., et al. (2005). Delayed-shot 3D depth migration. Geophysics 70 (6), E21–E28. doi:10.1190/1.2057980
Keywords: least-squares reverse time migration, inverse Hessian, fast imaging, generative adversarial network, deep learning
Citation: Huang Y, Huang J and Ma Y (2022) A fast least-squares reverse time migration method using cycle-consistent generative adversarial network. Front. Earth Sci. 10:967828. doi: 10.3389/feart.2022.967828
Received: 13 June 2022; Accepted: 28 June 2022;
Published: 24 August 2022.
Edited by:
Mourad Bezzeghoud, Universidade de Évora, PortugalReviewed by:
Bingluo Gu, China University of Petroleum, ChinaGang Yao, China University of Petroleum, Beijing, China
Qiancheng Liu, Institute of Geology and Geophysics (CAS), China
Copyright © 2022 Huang, Huang and Ma. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jianping Huang, jphuang@upc.edu.cn