Abstract
With high imaging accuracy, high signal-to-noise ratio, and good amplitude balance, least-squares reverse time migration (LSRTM) is an imaging algorithm suitable for deep high-precision oil and gas exploration. However, the computational costs limit its large-scale industrial application. The difference between traditional reverse time migration (RTM) and LSRTM is whether to eliminate the effect of the Hessian operator or not while solving Hessian matrix explicitly or eliminating the effect of the Hessian matrix implicitly has a very high requirement on computation or storage capacity. We simulate the inverse Hessian by training a cycle-consistent generative adversarial network (cycleGAN) to construct a mapping relationship between the RTM results and the true reflectivity models. The trained network is directly applied to the RTM imaging results, which improves the imaging quality while significantly reducing the calculation time. We select three velocity models and two velocity models respectively to generate the training and validation data sets, where the validation data is not involved in the training process. The prediction results on the validation data sets show that the trained network significantly improves the imaging quality with almost no additional in computational effort. Finally, we apply the network trained with only synthetics to the field data. The predicted results confirm the effectiveness and good generalization of the proposed method.
Introduction
For the high-precision imaging problem of complex geological formations and deep reservoirs, the reverse time migration (RTM) based on the two-way wave-equation is considered to be a better imaging method (Baysal et al., 1983; Chang, 1987). Essentially, RTM is only the conjugate transpose of the forward operator, not its inverse (Lailly, 1983). Therefore, influenced by the limited acquisition aperture, complex subsurface structure, and limited seismic wave bandwidth, the RTM results suffer from migration noise, non-uniform illumination, and insufficient resolution (Cheng et al., 2021), which prevents it from imaging deep and unconventional oil and gas reservoirs finely. In addition, conventional RTM can only image the subsurface discontinuities while the accurate amplitude information of the real formation lithology changes can not be guaranteed (Chen and Zhang, 2012).
To eliminate the effect of the Hessian matrix, the imaging problem of RTM can be solved as a least-squares inverse problem to suppress the migration noise, increase the resolution, and improve the amplitude fidelity, which is the least-squares reverse time migration (Tarantola, 1987; Chavent and Plessix, 1999; Nemeth et al., 1999; Valenciano et al., 2006). The least-squares reverse time migration (LSRTM) can be solved in model space (Hu et al., 2001; Rickett, 2003; Guitton, 2004) or in data space (Tarantola, 1987; Schuster, 1993; Chavent and Plessix, 1999; Nemeth et al., 1999). The LSRTM in model space requires solving the Hessian matrix explicitly and using its inverse to deblurring the RTM imaging results. Directly solving the full Hessian matrix in practical applications is extremely demanding in terms of computation and storage (Tang, 2009). Therefore, some researchers have proposed that the Hessian matrix can be approximated as a diagonal matrix under the assumption of high-frequency asymptotics and infinite aperture (Beylkin, 1985; Chavent and Plessix, 1999; Plessix and Mulder, 2004; Symes, 2008). However, this approximation is not suitable for the limited frequency range and acquisition geometry, and the Hessian matrix in this case is no longer diagonally dominant (Pratt et al., 1998; Chavent and Plessix, 1999; Plessix and Mulder, 2004; Valenciano et al., 2006). In regions below the high-velocity body or salt dome where illumination is lacking, the energy on the diagonal of the Hessian matrix is scattered along the non-diagonal direction (Albertin et al., 2004; Valenciano, 2008). Therefore, the diagonal Hessian matrix is of limited use for deblurring the migration imaging results, especially in regions with severe energy attenuation of seismic waves. Some non-diagonal elements of the Hessian matrix can be properly computed to compensate for the illumination unevenness in regions with severe absorption attenuation and enhance the deblurring effect (Albertin et al., 2004; Valenciano, 2008). However, solving the non-diagonal elements of the matrix will significantly increase the computational effort.
The LSRTM in data space is used to find the reflectivity model that best matches the observed seismic data by iterative algorithms such as the steepest descent method and the conjugate gradient method (Nemeth et al., 1999). Instead of solving the huge Hessian matrix directly, this approach implicitly removes the effect of the Hessian matrix by iteratively updating the reflectivity model. However, iterative solving causes a large amount of computation, and the convergence rate depends on suitable preconditioning constraints (Tang, 2009; Wu et al., 2021; Wu et al., 2022; Yao et al., 2022). To improve the computational efficiency of LSRTM, one approach is to compress the massive shot data into several super gathers to reduce the computational effort, which is the multisource migration method (Romero et al., 2000). To suppress the multi-source crosstalks, some researchers proposed the shot encoding techniques, such as amplitude encoding (Hu et al., 2016), phase encoding (Schuster et al., 2011; Li et al., 2014; Li et al., 2017) and plane wave encoding (Zhang et al., 2005; Dai and Schuster, 2013; Li et al., 2018). These schemes reduce the correlation of crosstalk noise in the imaging results of different super gathers by constructing suitable encoding functions and finally stack the imaging results to suppress the noise (Etgen, 2005; Zhang et al., 2005). However, when the observation system is not fixed, the source wavefield of the multisource forward simulation and the super gather synthesized from the observation data cannot be matched, resulting in slow convergence (Li et al., 2018). Liu and Peter (2018) propose a one-step LSRTM method to design a deblurring preconditioner using the Wiener filter, reducing the computational effort to twice the RTM. In summary, LSRTM in model space requires certain calculations and significant storage costs to solve the Hessian matrix explicitly, while LSRTM in data space requires a large amount of computation to solve the reflectivity iteratively and does not require much storage.
In recent years, with the popularity of artificial intelligence, many researchers introduced deep learning techniques to solve geophysical problems such as fault segmentation (Wu et al., 2019), seismic data denoising (Yu et al., 2019; Saad and Chen, 2020), and first-arrival picking (Hu et al., 2019; Duan and Zhang, 2020). These solutions use techniques such as DNN, CNN, RNN, etc. These networks can be broadly classified into two categories based on the application scenarios: discriminative models that solve classification, identification, or localization problems, and generative models that solve the mapping relationship between data from two different domains. The generative adversarial network (GAN) combines the advantages of both types of networks by using a generative model to produce data from another domain and then using a discriminative model to evaluate the generated data (Goodfellow et al., 2014). Through adversarial training, GAN achieves state-of-the-art in solving the problem of style transfer between data from different domains. The cycleGAN (Zhu et al., 2017) consists of two GANs that ensure the structural consistency of the input and output data through cyclic consistency, taking full advantage of the mapping relationship between the source and target domain data. This network structure is well suitable for transferring the RTM data into a true reflectivity model and eliminating the blurring effect of Hessian matrix.
In this study, we propose to approximate the inverse Hessian by using cycleGAN for the deblurring purpose. The cycleGAN is trained by generating data sets of RTM imaging results with reflectivity models. The trained network is directly applied to the migration images to output high-quality imaging profiles. The training stage of the network consumes a certain computational effort and has an almost negligible computational cost. The proposed method is validated on both synthetic data and field data.
Methodology
Migration theory
For a given reflectivity model and a forward modeling operator , the observed data can be expressed as:
The goal of seismic imaging is to find the exact to fit the observed data . Conventional reverse time migration can be expressed as:where is the matrix representation of the reverse time migration imaging result and is the matrix representation of the RTM operator. The is only the transpose of the forward operator, not its inverse, and cannot completely reverse the seismic wave propagation, so the migrated images of reverse time migration do not accurately represent the true reflectivity model.
The solution of the true reflectivity model can be solved in a least-squares framework. We define the misfit function as follows:
Applying the transpose forward operator to both sides of Eq. 1, Eq. 4 is derived as follows:
Therefore, the precise least-squares solution of the reflectivity model can be obtained by:where is the inverse of the Hessian matrix, and is the images of the reverse time migration. The relationship between the RTM imaging result and the true reflectivity model can be expressed as:
Analysis of Hessian matrix
The expression of the conventional two-dimensional acoustic wave equation is:where is the slowness field of the medium, is the corresponding velocity field, and is the representation of the source wavelet in the frequency domain, indicating that the source point is at and the angular frequency is . Based on the assumption that the inverse problem is approximately linear, the expression of the Hessian matrix for this two-dimensional acoustic wave equation can be written in the following form:where is the Green’s function at the source side, is the Green’s function at the receiver side, and is the source signature. The Green’s function term reflects the illumination of the seismic wave during migration, while the source wavelet term has an impact on the imaging resolution.
According to Eq. 6, it can be seen that the RTM results are missing the deblurring filtering of the inverse Hessian matrix with respect to the real reflectivity models. And according to Eq. 8, it can be concluded that the wavelet term and the Green’s function term in the expression of the Hessian matrix affect the imaging resolution and wavefield propagation, respectively. Therefore, the migrated images of RTM exhibit low imaging resolution, uneven illumination, and migration noise due to the lack of the inverse Hessian filtering. By solving the inverse Hessian matrix and applying it to the imaging results of the RTM, the true reflectivity model can be obtained, which is the least-squares reverse time migration in the model space. Theoretically, the LSRTM in model space only requires a certain computational cost to compute the inverse of the Hessian matrix in advance, and then the true reflectivity model can be acquired quickly. However, it is shown that solving the Hessian matrix requires storing the Green’s function, which is very large in practical applications, especially for 3D seismic data. Another approach is the LSRTM in the data space, which is an iterative solution to continuously update the reflectivity model so that the error between the simulated and observed data reaches the desired value. The LSRTM in the data domain does not take into account the storage cost, but it is computationally intensive, about 2N times more than the conventional migration (N denotes the number of iterations). Moreover, the convergence of the LSRTM in data space is very slow without proper gradient preconditioning.
Our goal is to approximate the inverse of the Hessian matrix at a small cost. In the proposed method, we train cycleGAN to simulate the inverse Hessian and apply the trained network to process the RTM imaging data and play the role of the deblurring operator. First, we select some velocity models and generate the true reflectivity model . Through finite-difference forward simulation, we can obtain the synthetic observation data . Then, the RTM imaging of the observation data yields the migrated images after being filtered by the Hessian blurring operator. The source domain data and the target domain data are fed into the network for training. According to Eq. 6, the trained network is the approximate inverse Hessian. After training, we apply the network to the new RTM imaging data to quickly obtain the least-squares solution in Eq. 5.
Network and loss function
The architecture of cycleGAN
GAN was proposed by Goodfellow et al. (2014), and it has good applicability to the problem of the source-to-target domain transformation. Conventional neural networks are used to find an optimal mapping relationship from input to output by backpropagation algorithms. The nature of the network can be viewed as a complex higher-order polynomial. GAN generally consists of two parts of the network, one as a generator and the other as a discriminator. During the training process of the network, the generator acts as a regular neural network, converting the input into data similar to the label, while the discriminator takes the output data of the generator as input and tries to distinguish the generated data from the real data in the training sets. Through this adversarial training, both the generator and the discriminator are continuously optimized, and eventually, we get the desired generator as a prediction network.
CycleGAN consists of 2 mirror-symmetric networks and , which together form a cycle network. As shown in Figure 1, cycleGAN has 2 discriminators and , and shares two generators and . A sample from the data distribution in the source domain is input to the first network , and fake and cyclic are produced by generators and in turn, then, fake enters the discriminator to be identified as true or false. Similarly, for any sample in the data distribution in the target domain, it can also be used as input to train the second network . The generator consists of an encoder, a converter, and a decoder. As shown in Figure 2, the encoder-decoder structure is similar to most convolutional neural networks, with convolutional and deconvolutional layers for feature extraction and reduction, and the converter structure is implemented by a multilayer Resnet. The discriminator is presented in Figure 3, which is implemented by a basic convolutional neural network, whose discriminating image requires extracting features from the input and finally adding a one-dimensional output convolutional layer to determine whether the extracted features belong to a specific class.
FIGURE 1

The architecture of cycleGAN.
FIGURE 2

The specific structure of the generator, consists of 2 convolutional layers forming the encoder, 7 Resnet blocks as the converter, and 2 deconvolutional layers as the decoder.
FIGURE 3

The detailed architecture of the discriminator, 4 convolutional layers are used to extract the features of the input data, and finally, a convolutional layer is added to output the discriminant score.
Loss function analysis
The loss function is a measure of how good the output of the network is, and it is also the basis for correcting the tensor gradient in the backpropagation process. For a given size training batch, assuming the data distribution of the source domain is and the data distribution of the target domain is , the data distribution of each training data and can be denoted as and . The mathematical expectation of the training data can be expressed as and . Therefore, the objective function of a conventional GAN can be expressed as (Goodfellow et al., 2014; Benaim and Wolf, 2017; Zhu et al., 2017):
During training, the generator tries to generate that is more similar to the reflectivity model, while the discriminator tries to distinguish the generated from the real reflectivity model in the target domain. Thus, the goal of the generator is to minimize the objective function, while the discriminator tries to maximize it. For the first GAN, the final goal is , and similarly, for the other GAN, its target is .
By the above constraint of the objective function, GAN can theoretically map any data from the source domain to the target domain. However, only this constraint may result in the generator just generating a result with reflectivity characteristics, while the structural content is not the same as the source domain data. That is, only style transfer is achieved by training, and there is no guarantee that the content does not change. For any source domain data , we need to guarantee , which is also called forward cycle consistency. Similarly, for any target domain data , we need it to satisfy the backward cycle consistency: . Therefore, cycleGAN introduces a cyclic consistency error function as follows (Zhu et al., 2017):With the above loss function constraint, any data in the source or target domain can be effectively restored to the original data features in the end after two generator cycles. However, when the network is trained for application, the situation that the style completes the migration but the content changes still occurs. The analysis shows that only one generator is used as the prediction network when applied, and the cyclic consistency error function does not impose constraints on the intermediate variables. Therefore, we introduce the identity loss function to constrain the intermediate variables, and its expression is (Taigman et al., 2016; Benaim and Wolf, 2017; Zhu et al., 2017):
Therefore, the full loss function when training the cycleGAN is shown as follows:
Data sets preparation and training
Training data sets production
As shown in Figures 4A–C, the Sigsbee2A model, SEAM model, and Pluto model are selected as the base models to produce the training data sets. The migrated images corresponding to the three models are generated by 2-8 finite-difference forward simulation and reverse time migration as shown in Figures 4D–F. Meanwhile, we produce the reflectivity model based on the velocity model as shown in Figures 4H–J. Considering the large band range variability between the reflectivity model and the migration imaging results, we adopt the convolution approach and conduct it with the reflectivity and the Ricker wavelet as the target domain data. Finally, we apply data enhancement such as cropping and resampling to generate 300 pairs of training data. Some of the training data are shown in Figure 5. The first row shows the RTM imaging data in the source domain, and the second row shows the convolution results in the target domain.
FIGURE 4

Three velocity models for producing training data sets in the source and target domains. (A) Sigsbee2A velocity model, (B) SEAM velocity model, and (C) Pluto velocity model. The corresponding RTM imaging results (D–F) and reflectivity models (H–J) for making the input and output samples respectively.
FIGURE 5

Four pairs of representative samples from the training data sets. The first row (A–D) are the migrated images in the source domain, and the second row (E–H) denotes the reflectivity in the target domain.
Training process
For , the training first fixes the generator and trains the discriminator separately to improve the discriminator’s ability to identify the output results of the generator, and then fixes the discriminator and trains the generator individually to generate more realistic results and bypass the discriminator’s recognition. With this adversarial training approach, the ability of the generator to map to the target domain can be greatly improved. Similarly, for the other , such adversarial training is also used. When the two GANs are trained for one epoch, the values of all loss functions are accumulated for backpropagation. Finally, the generator is used as the expected prediction network for approximating the inverse Hessian matrix.
The number of epochs of the network training is set to 200 with a learning rate of 0.0002, which gradually decreased after the 100th epoch. The network is weighted with a cyclic consistency loss of 10 and the weight of the identity loss is set to 5. The training is based on the PyTorch platform, using a single NVIDIA RTX5000 GPU, and the total training time is 10 h.
Numerical tests on validation sets
BP 2.5D model
Figure 6A shows the BP2.5D velocity model with the reflectivity model shown in Figure 6B. We perform a finite-difference forward simulation of the velocity model in Figure 6A, then smooth the velocity model into migration velocity and perform reverse time migration to generate the imaging profile in Figure 6C. There are obvious low-frequency noise and acquisition footprints in the migrated image of the RTM, and the imaging energy is discontinuous in the region of violent tectonic undulations, meanwhile, the illumination is uneven on both sides of the model. We input the imaging result of Figure 6C into the trained network, and the obtained prediction result is shown in Figure 6D. The imaging quality in the predicted result is significantly improved, the surface acquisition footprints are eliminated very cleanly, and the low-frequency noise is well removed. The trained network compensates well for the uneven illumination in the migration result, and the overall imaging energy distribution is relatively uniform. It can be noted that the network is not completely accurate in predicting the result at the depth of 2 KM on the left side of the model since the imaging energy of the original RTM profile is very heterogeneous, resulting in the network not being able to predict it well. Figure 7 shows an enlarged view of the red dashed boxed area in Figure 6, with the first column showing the RTM imaging results, the second column denoting the network prediction results, and the third column representing the true reflectivity models. Comparing the first set of results (Figures 7A–C), it can be seen that the proposed method can better play the role of the deblurring operator and significantly improve the imaging quality. The second set of results (Figures 7D–F) shows that the trained network can not fully reduce the image of the RTM to the true reflectivity model, and it relies to some extent on the structure information of the original migration image, which differs from the real reflectivity model in some details, but generally maintains the consistency with the real reflectivity model.
FIGURE 6

BP 2.5D velocity model (A) and the corresponding reflectivity model (B). Input the RTM imaging result (C) into the well-trained network, the prediction result is shown in (D).
FIGURE 7

The zoom views of the red dashed boxed area in Figure 6. The first column represents the RTM results (A,D), the second column denotes the predictions (B,E) of the network and the third column is the true reflectivity model (C,F).
Marmousi model
Figure 8A shows the Marmousi velocity model, and Figure 8B shows its corresponding reflectivity model. The imaging result obtained after finite-difference simulation and reverse time migration of Figure 8A is shown in Figure 8C. The migrated image of RTM has more low-frequency noise in the shallow layer due to the blurring filtering of Hessian matrix. At the same time, the imaging resolution of the model bottom structure is low due to the influence of the deep high-speed layer shielding. Figure 8D shows the result of network prediction. Compared with Figure 8C, it can be concluded that the global resolution of the predicted imaging result is improved, and the reflection layer is finer. Most of the low-frequency noise in the shallow layer is removed, with a small amount of residual on both sides, and the trained network is also very clean for removing the acquisition footprints in the profile. The imaging quality of the deeper layers is also improved, however, some artifacts appear since some stray imaging energies in the result of the RTM are mistaken by the network as layers and recovered. Figure 9 shows an enlarged view of the red dashed boxed area in Figure 8. Figure 9A shows a shallow zoom of the RTM imaging result, where strong low-frequency noise and acquisition footprints exist within 0.5 KM of depth. These clutter energies are barely visible in the predicted result in Figure 9B, and the imaging energies at the interface are more converged. Comparing the second set of zooming results, it can be seen that the resolution of the result predicted by the network has improved significantly and the overall imaging energy is more balanced.
FIGURE 8

The Marmousi velocity model (A) and the corresponding reflectivity model (B). Input the RTM imaging result (C) into the well-trained network, the prediction result is shown in (D).
FIGURE 9

The zoom views of the red dashed boxed area in Figure 8. The first column represents the RTM results (A,D), the second column denotes the predictions (B,E) of the network and the third column is the true reflectivity model (C,F).
Field data test
We test the effectiveness of the proposed method with field data. We take the velocity model in Figure 10A as the migration velocity and perform the reverse time migration to obtain the imaging result shown in Figure 10B. We approximate the deblurring operator with the network previously trained on synthetic data, which acts directly on the migrated image of the field data. We input the migration result in Figure 10B into the network, and the prediction result output by the generator is shown in Figure 10C. Compared with Figure 10B, there is a significant improvement in the resolution of the network prediction result, and a large amount of migration noise in the RTM image is well removed. Without damaging the effective signal, the surface acquisition footprints are eliminated very cleanly, and the shallow reflection layer is more obvious with more imaging details. However, the network does not handle the deep high steep structure very well. This is because the reflectivity model of our constructions is calculated along the depth direction. When the discretized model is small or not adequately sampled, its structure has poor continuity in the lateral direction. In contrast, the seismic wavefield propagates in all directions, and the imaging results have good continuity even on small models. Therefore, this results in an incompletely accurate mapping relationship between source-domain data and target-domain data, which affects the prediction results of the network.
FIGURE 10

(A) migration velocity model, (B) RTM imaging result, and (C) the prediction result of the network trained only by synthetic data sets.
Conclusion
We propose a fast least-squares reverse time migration method based on cycleGAN. In the proposed method, cycleGAN is used to approximate the inverse Hessian matrix to play the role of a deblurring operator that acts directly on the RTM imaging results. By constructing a training data sets consisting of source domain data (migration imaging results) and target domain data (true reflectivity model), two GANs are trained separately from the two domains, and a generator mapping the source domain to the target domain is taken as the final prediction network after adversarial training. Synthetic data tests demonstrate the effectiveness of the proposed method in removing low-frequency noise, improving resolution, balancing illumination, and compensating for deep imaging energy. Applying the network trained only with synthetics directly to the field data, the prediction result shows that the proposed method has good generalization and good applicability to data sets with large variability. One point worth noting is that in the prediction result of the field data, the recovery of high steepness structures by the method is not very well. This problem can be solved by optimizing the label-making scheme and reducing the spatial sampling interval. Compared with the massive storage requirement of the Hessian matrix by the LSRTM in the model space, the storage demands of the proposed method are not high. Compared with the LSRTM in data space, the proposed method only consumes certain computational resources in the training stage and requires almost no computational costs in the prediction process.
Statements
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
YH and JH contributed to the idea and methodology, YH was responsible for the training data production, model and field data testing, and the writing of this manuscript, and YM checked and polished the manuscript. All authors have read the manuscript and agreed to publish it.
Funding
This research is supported by the National Key R&D Program of China (no. 2019YFC0605503), the National Outstanding Youth Science Foundation (no. 41922028), and the Major Scientific and Technological Projects of CNPC (no. ZD 2019-183-003).
Acknowledgments
The authors are grateful to the editors and reviewers for their review of this manuscript. Meanwhile, the authors would like to thank the PyTorch platform for providing excellent deep learning tools.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1
AlbertinU.YingstD.KitchensideP.TcheverdaV. (2004). Expanded Abstracts: SEG, 949–952.True-amplitude beam migration: 74st annual international meeting
2
BaysalE.KosloffD. D.SherwoodJ. W. C. (1983). Reverse time migration. Geophysics48 (11), 1514–1524. 10.1190/1.1441434
3
BenaimS.WolfL. (2017). One-sided unsupervised domain mapping. Adv. Neural Inf. Process. Syst., 752–762.
4
BeylkinG. (1985). Imaging of discontinuities in the inverse scattering problem by inversion of a causal generalized Radon transform. J. Math. Phys.26 (1), 99–108. 10.1063/1.526755
5
ChangW. F.McMechanG. A. (1987). Elastic reverse‐time migration. Geophysics52 (3), 1365–1375. 10.1190/1.1442249
6
ChaventG.PlessixR. E. (1999). An optimal true‐amplitude least‐squares prestack depth‐migration operator. Geophysics64 (2), 508–515. 10.1190/1.1444557
7
ChenS. C.ZhangB. (2012). Steeply-dipping structures migration based on the wavefield vertical- and horizontal-extrapolation. Chinese J. Geophys.55 (4), 1300–1306.
8
ChengJ. B.WangH. Z.MaZ. T. (2021). Pre-stack depth migration with finite-difference method in frequency-space domain. Chinese J. Geophys.44 (3), 389–395.
9
DaiW.SchusterG. T. (2013). Plane-wave least-squares reverse-time migration. Geophysics78 (4), S165–S177. 10.1190/geo2012-0377.1
10
DuanX. D.ZhangJ. (2020). Multitrace first-break picking using an integrated seismic and machine learning method. Geophysics85 (4), WA269–WA277. 10.1190/geo2019-0422.1
11
EtgenJ. T. (2005). Expanded Abstracts: SEG.How many angles do we really need for delayed-shot migration? 75st annual international meeting
12
Gerhard PrattR. G.ShinC.HicksG. J. (1998). Gauss-Newton and full Newton methods in frequency-space seismic waveform inversion. Geophys. J. Int.133 (2), 341–362. 10.1046/j.1365-246x.1998.00498.x
13
GoodfellowI.Pouget-AbadieJ.MirzaM.XuB.Warde-FarleyD.OzairS.et al (2014). Generative adversarial nets. Adv. Neural Inf. Process. Syst., 2672–2680.
14
GuittonA. (2004). Amplitude and kinematic corrections of migrated images for nonunitary imaging operators. Geophysics69 (4), 1017–1024. 10.1190/1.1778244
15
HuJ. G.WangH. Z.FangZ. Y.LiT. C.ZhangJ. N. (2016). Efficient amplitude encoding least-squares reverse time migration using cosine basis. Geophys. Prospect.64 (6), 1483–1497. 10.1111/1365-2478.12356
16
HuJ.SchusterG. T.ValasekP. A. (2001). Poststack migration deconvolution. Geophysics66 (3), 939–952. 10.1190/1.1444984
17
HuL. L.ZhengX. D.DuanY. T.YanX. F.HuY.ZhangX. L.et al (2019). First-arrival picking with a U-net convolutional network. Geophysics84 (6), U45–U57. 10.1190/geo2018-0688.1
18
LaillyP. (1983). The seismic inverse problem as a sequence of before stack migrations. Conf. Inverse Scatt. Theory Appl.15 (2), 206–220.
19
LiC.HuangJ. P.LiZ. C.WangR. R. (2018). Plane-wave least-squares reverse time migration with a preconditioned stochastic conjugate gradient method. Geophysics83 (1), S33–S46. 10.1190/geo2017-0339.1
20
LiC.HuangJ. P.LiZ. C.WangR. R. (2017). Preconditioned prestack plane-wave least squares reverse time migration with singular spectrum constraint. Appl. Geophys.14 (1), 73–86. 10.1007/s11770-017-0599-8
21
LiC.HuangJ. P.LiZ. C.WangR. R.SunM. M. (2018). Least-squares reverse time migration with a hybrid stochastic conjugate gradient method. Oil Geophys. Prospect.53 (6), 1188–1197.
22
LiZ. C.GuoZ. B.TianK. (2014). Least-squares reverse time migration in visco-scoustic media. Chin. J. Geophys.57 (1), 214–228.
23
LiuQ. C.PeterD. (2018). One-step data-domain least-squares reverse time migration. Geophysics83 (4), R361–R368. 10.1190/geo2017-0622.1
24
NemethT.WuC.SchusterG. T. (1999). Least‐squares migration of incomplete reflection data. Geophysics64 (1), 208–221. 10.1190/1.1444517
25
PlessixR. E.MulderW. A. (2004). Frequency-domain finite-difference amplitude-preserving migration. Geophys. J. Int.157 (3), 975–987. 10.1111/j.1365-246x.2004.02282.x
26
RickettJ. E. (2003). Illumination‐based normalization for wave‐equation depth migration. Geophysics68 (4), 1371–1379. 10.1190/1.1598130
27
RomeroL. A.GhigliaD. C.OberC. C.MortonS. A. (2000). Phase encoding of shot records in prestack migration. Geophysics65 (2), 426–436. 10.1190/1.1444737
28
SaadO. M.ChenY. K. (2020). Deep denoising autoencoder for seismic random noise attenuation. Geophysics85 (4), V367–V376. 10.1190/geo2019-0468.1
29
SchusterG. T. (1993). Expanded Abstracts: SEG, 110–113.Least-squares cross-well migration: 63st annual international meeting
30
SchusterG. T.WangX.HuangY.DaiW.BoonyasiriwatC. (2011). Theory of multisource crosstalk reduction by phase-encoded statics. Geophys. J. Int.184 (3), 1289–1303. 10.1111/j.1365-246x.2010.04906.x
31
SymesW. W. (2008). Approximate linearized inversion by optimal scaling of prestack depth migration. Geophysics73 (2), R23–R35. 10.1190/1.2836323
32
TaigmanY.PolyakA.WolfL. (2016). Unsupervised cross-domain image generation: arXiv preprint arXiv:1611.02200.
33
TangY. (2009). Target-oriented wave-equation least-squares migration/inversion with phase-encoded Hessian. Geophysics74 (6), WCA95–WCA107. 10.1190/1.3204768
34
TarantolaA. (1987). Inverse problem theory. Methods for data fitting and model parameter estimation. Elsevier.
35
ValencianoA. A.BiondiB.GuittonA. (2006). Target-oriented wave-equation inversion. Geophysics71 (4), A35–A38. 10.1190/1.2213359
36
ValencianoA. (2008). Imaging by wave-equation inversion: Ph.D. dissertation. Stanford University.
37
WuB.YaoG.CaoJ. J.WuD.LiX.LiuN. C.et al (2022). Huber inversion-based reverse-time migration with de-primary imaging condition and curvelet-domain sparse constraint. Petroleum Sci.18. 10.1016/j.petsci.2022.03.004
38
WuD.WangY. H.CaoJ. J.da SilvaN. V.YaoG. (2021). Least-squares reverse-time migration with sparsity constraints. J. Geophys. Eng.18, 304–316. 10.1093/jge/gxab015
39
WuX. M.LiangL. M.ShiY. Z.FomelS. (2019). FaultSeg3D: Using synthetic data sets to train an end-to-end convolutional neural network for 3D seismic fault segmentation. Geophysics84 (3), IM35–IM45. 10.1190/geo2018-0646.1
40
YaoG.WuB.da SilvaN. V.DebensH. A.WuD.CaoJ. J.et al (2022). Least-squares reverse time migration with a multiplicative Cauchy constraint. Geophysics87 (3), S151–S167. 10.1190/geo2021-0183.1
41
YuS. W.MaJ. W.WangW. L. (2019). Deep learning for denoising. Geophysics84 (6), V333–V350. 10.1190/geo2018-0668.1
42
ZhangY.SunJ.NotforsC.GrayS.ChernisL.YoungJ.et al (2005). Delayed-shot 3D depth migration. Geophysics70 (6), E21–E28. 10.1190/1.2057980
43
ZhuJ. Y.ParkT.IsolaP.EfrosA. A. (2017). Unpaired image-to-image translation using cycle-consistent adversarial networks, IEEE International Conference on Computer Vision (ICCV). IEEE, 2242–2251. 10.1109/iccv.2017.244
Summary
Keywords
least-squares reverse time migration, inverse Hessian, fast imaging, generative adversarial network, deep learning
Citation
Huang Y, Huang J and Ma Y (2022) A fast least-squares reverse time migration method using cycle-consistent generative adversarial network. Front. Earth Sci. 10:967828. doi: 10.3389/feart.2022.967828
Received
13 June 2022
Accepted
28 June 2022
Published
24 August 2022
Volume
10 - 2022
Edited by
Mourad Bezzeghoud, Universidade de Évora, Portugal
Reviewed by
Bingluo Gu, China University of Petroleum, China
Gang Yao, China University of Petroleum, Beijing, China
Qiancheng Liu, Institute of Geology and Geophysics (CAS), China
Updates
Copyright
© 2022 Huang, Huang and Ma.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jianping Huang, jphuang@upc.edu.cn
This article was submitted to Solid Earth Geophysics, a section of the journal Frontiers in Earth Science
Disclaimer
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.