- 1Zhejiang Key Laboratory of Digital Precision Measurement Technology Research, Hangzhou, China
- 2Advanced Manufacturing Metrology Research Center, Zhejiang Institute of Quality Sciences, Hangzhou, China
- 3Key Laboratory of Acoustics and Vibration Applied Measuring Technology, State Administration for Market Regulation, Hangzhou, China
Reliable visual perception is essential for autonomous driving test scenario generation, yet adverse weather and lighting variations pose significant challenges to simulation robustness and generalization. Traditional unpaired image-to-image translation methods primarily rely on RGB-based transformations, often resulting in geometric distortions and loss of structural consistency, which can negatively impact the realism and accuracy of generated test scenarios. To address these limitations, we propose a Depth-Aware Dual-Branch Generative Adversarial Network (DAB-GAN) that explicitly incorporates depth information to preserve spatial structures during scenario generation. The dual-branch generator processes both RGB and depth inputs, ensuring geometric fidelity, while a self-attention mechanism enhances spatial dependencies and local detail refinement. This enables the creation of realistic and structure-preserving test environments that are crucial for evaluating autonomous driving perception systems, especially under adverse weather conditions. Experimental results demonstrate that DAB-GAN outperforms existing unpaired image-to-image translation methods, achieving superior visual fidelity and maintaining depth-aware structural integrity. This approach provides a robust framework for generating diverse and challenging test scenarios, enhancing the development and validation of autonomous driving systems under various real-world conditions.
1 Introduction
Realistic and diverse test scenario generation is essential for evaluating autonomous driving systems, as perception models must effectively operate in complex and dynamic environments (Cai J. et al., 2024; Li X. et al., 2023; Zhang J. W. et al., 2024). This challenge becomes even more pronounced in adverse weather conditions, such as fog, rain, and low-light scenarios, which significantly impact object recognition, depth estimation, and overall scene understanding. These factors make it particularly difficult to assess the robustness and generalization ability of perception models, highlighting the need for realistic and varied test scenarios to ensure comprehensive evaluation.
Despite the importance of real-world testing, collecting diverse adverse weather data remains costly, time-consuming, and logistically challenging. Moreover, real-world datasets often suffer from imbalance and limited coverage of extreme conditions, restricting their utility in comprehensive validation and robustness assessment (Agarwal et al., 2025; Lan et al., 2024; Li Y. et al., 2023). As a result, simulation-based testing has become a critical tool for autonomous driving development, allowing for the controlled generation of challenging environmental conditions to enhance model reliability and adaptability (Biagiola and Tonella, 2024; Huang et al., 2025; Sadid and Antoniou, 2024). A key requirement for effective simulation is the ability to generate photorealistic and geometrically consistent test scenarios that accurately reflect real-world conditions.
Unpaired image-to-image translation has emerged as a powerful approach for scenario generation and domain adaptation, enabling the transformation of clear-weather images into adverse weather conditions without requiring paired training data (Ding et al., 2024; Wei et al., 2024). This technique is particularly valuable for augmenting training datasets and improving model robustness (Ye M. et al., 2024; Ye R. et al., 2024). However, traditional unpaired image-to-image translation methods primarily rely on RGB-based transformations, often resulting in geometric distortions and loss of structural consistency due to their pixel-wise nature, limiting their effectiveness for realistic autonomous driving test scenario generation.
Geometric consistency constitutes a fundamental principle for enhancing visual fidelity in autonomous driving simulation environments. In the domain of scene perception, the integration of depth information, edge-aware modules, and attention mechanisms has been widely recognized as an effective means of improving the performance of generative models (Tong et al., 2025a,b; Tong et al., 2024). Tong et al. (2025a) introduced an ORB-assisted sparse optical flow constraint as auxiliary guidance, which enhances the reliability of depth estimation in uncertain regions and significantly improves SLAM performance. Chen M. et al. (2024) proposed a novel framework that leverages depth distribution priors to regulate cross-domain mixing strategies and employs a multi-task Transformer-based fusion mechanism to enhance structural awareness, effectively mitigating scene layout inconsistencies in unsupervised domain adaptation. Collectively, these approaches underscore the importance of structural priors in geometry-aware modeling and provide valuable insights into addressing the limitations of CycleGAN in realistic scene generation for autonomous driving applications.
Inspired by the aforementioned studies, we explore the incorporation of depth information and attention mechanisms into the CycleGAN framework, we propose a Depth-Aware Dual-Branch Generative Adversarial Network (DAB-GAN) that explicitly incorporates depth information into the scenario generation process. By leveraging a dual-branch generator that simultaneously processes RGB and depth inputs, our approach enhances geometric fidelity and structural consistency in translated images. Additionally, a self-attention mechanism is integrated to refine spatial dependencies and local details, allowing for more realistic and structure-preserving transformations across diverse environmental conditions.
The contributions of this work are threefold:
1. Depth-Aware Dual-Branch Architecture: We introduce a novel generative model that processes RGB and depth information in parallel, ensuring geometric consistency during image translation.
2. Self-Attention Mechanism: By incorporating self-attention, our model enhances spatial coherence and improves the realism of translated images, particularly in adverse weather scenarios.
3. Superior Performance in Autonomous Driving Scenario Generation: Experimental results demonstrate that DAB-GAN surpasses existing unpaired image-to-image translation methods, achieving higher visual fidelity and preserving depth-aware structural integrity across diverse driving environments.
The structure of this paper is as follows: the second section provides a comprehensive review of related research on Depth-Aware Autonomous Driving Test Scenario Generation. The third section presents a detailed description of the proposed method, highlighting its integration with the dual-branch architecture and self-attention mechanism. The fourth part reports the experimental results and performance evaluations, while the fifth part concludes the paper with key findings and potential directions for future research.
2 Related works
2.1 Generative adversarial networks for image translation
GANs (Goodfellow et al., 2014) have significantly advanced image synthesis and translation, broadly categorized into unconditional GANs and conditional GANs.
Unconditional GANs learn to generate images of a specific category solely from input images, without requiring explicit conditional guidance. Radford et al. (2016) introduced convolutional architectures that improved training stability and image quality, while StyleGAN (Karras et al., 2020, 2021a,b) further refined image synthesis through a style-based architecture, enabling fine-grained control over image attributes and facilitating high-resolution image generation.
Conditional GANs, in contrast, leverage auxiliary information to guide the image generation process and can be further classified into paired image-to-image translation and unpaired image-to-image translation (Chen L. et al., 2024; Shubham et al., 2024). While paired image-to-image translation has demonstrated significant advancements across various applications, its reliance on paired training data poses a major challenge due to the high cost and limited availability of such datasets in real-world scenarios (Liu M. et al., 2017; Shrivastava et al., 2017; Zhu et al., 2017). To address this limitation, unpaired image-to-image translation methods have been developed, enabling models to learn and transfer styles between different domains without the need for explicitly paired samples (Cai X. et al., 2024; Lin et al., 2025; Zhu et al., 2017). Liu M. et al. (2017) introduced CoGAN, which establishes a coupling relationship between generated images and style images, allowing them to interact and influence each other during training. CycleGAN (Zhu et al., 2017), on the other hand, proposed a cycle-consistency constraint, ensuring that an image translated from domain A to B can be accurately reconstructed back to A. This self-supervised learning framework effectively preserves structural and semantic consistency without the requirement for paired data, making CycleGAN a widely adopted baseline for unpaired image translation tasks (Liu et al., 2024; Yang L. et al., 2024; Ye H. et al., 2024).
Building on the success of CycleGAN, this work integrates depth information and self-attention mechanisms to further enhance structural preservation, semantic coherence, and visual fidelity in unpaired image-to-image translation, with a specific focus on autonomous driving test scenario generation. By leveraging these enhancements, our approach aims to improve the realism and consistency of generated test environments, facilitating more robust evaluations of autonomous driving perception systems under diverse conditions.
2.2 Depth map utilization in image synthesis
Depth information is pivotal in scene understanding, offering geometric constraints that enhance spatial consistency during image translation (Ming et al., 2021; Piccinelli et al., 2024). While depth-aware processing has been extensively explored in applications such as 3D reconstruction, depth estimation, and image super-resolution, its integration into unpaired image-to-image translation remains underexplored. Prior research has primarily focused on monocular depth estimation and stereo image synthesis, utilizing depth maps as auxiliary inputs to bolster structural consistency (Hong et al., 2024; Hong et al., 2022). However, within the realm of GAN-based translation, the incorporation of depth maps into the generative process is seldom addressed.
Recent studies have begun to explore depth-guided generative models for style transfer and domain adaptation. Liu X. et al. (2017) employed depth maps to preserve spatial layout during style transfer, treating depth as a supplementary constraint, and achieved promising results. Building upon this idea, Ioannou and Maddock (2022) utilized an advanced depth prediction network to incorporate depth preservation as an additional loss alongside content and style losses. Experiments demonstrated that their proposed method outperformed other models in comparative evaluations. Similarly, Hong et al. (2022) proposed DaGAN and its improved version, DaGAN++ (Hong et al., 2024), which integrate depth information to guide facial keypoint detection and cross-modal attention learning in talking head video generation. These works highlight the potential of incorporating depth into generative tasks. However, such approaches often treat depth information as a secondary feature rather than as a core component of the image transformation process.
In contrast, our proposed dual-branch GAN architecture explicitly processes depth maps alongside RGB images, enabling the model to jointly learn appearance-based style modifications and depth-preserving geometric structures.
2.3 Attention mechanisms in GANs
In traditional Generative Adversarial Networks (GANs), convolutional layers are inherently limited by their local receptive fields, making it challenging for the generator to simultaneously capture global structures and preserve fine-grained details. Attention mechanisms, on the other hand, have demonstrated remarkable effectiveness in enhancing image generation quality by capturing long-range dependencies and refining intricate details (Xue et al., 2024; Ye and Wang, 2024). Due to their ability to selectively focus on important regions of an image while maintaining contextual coherence, attention mechanisms have been widely adopted in GAN-based image synthesis, particularly for refining style transfer and enhancing content preservation.
Bakht et al. (2024) incorporated Multi-Level Attention into a GAN framework to improve underwater image restoration, effectively preserving textures, edges, and object structures. Experimental evaluations demonstrated the superior performance of their approach, further validating its efficacy in underwater scene analysis. Yildiz et al. (2024) integrated self-attention into a GAN-based single-image generation framework, with extensive evaluations across multiple datasets confirming significant improvements in image quality, training stability, and texture preservation, outperforming existing models. These studies highlight the transformative impact of attention mechanisms in image generation, particularly in scenarios requiring fine-detail preservation and structural coherence.
Given the demonstrated effectiveness of attention mechanisms in GAN-based image generation, we extend CycleGAN by incorporating self-attention into its generator network. This integration allows the model to dynamically prioritize spatially significant regions while maintaining global style consistency, thereby enhancing structural coherence, semantic integrity, and overall visual realism in unpaired image-to-image translation.
2.4 Image translation for autonomous driving test scenario generation
The generation of realistic and diverse test scenarios is essential for the development and validation of autonomous driving systems. Traditional methods, such as real-world driving and closed-course simulations, often fail to capture the full range of traffic situations and rare edge cases that autonomous vehicles may encounter (Li et al., 2022; Zhang J. W. et al., 2024). To address these limitations, image translation techniques have emerged as a promising solution for generating diverse and comprehensive test scenarios.
Image translation refers to methods that transform images from one domain to another, such as converting a clear road scene into foggy or nighttime conditions. This capability is particularly valuable for autonomous driving, as it enables the simulation of various environmental scenarios, such as different weather conditions and lighting, which can significantly enhance the testing of vehicle perception systems (Lambertenghi and Stocco, 2024). Lakmal et al. (2024) tackled the challenge of night-to-day image translation in autonomous driving, where acquiring paired night-day images in real-world settings is inherently difficult. To address this, they first employed CycleGAN to synthesize a low-quality paired dataset, which served as the training data for their proposed NtD-GAN model. This model demonstrated improved translation quality over existing approaches. However, its generalization capability remains limited, particularly under diverse and complex driving conditions such as rain, fog, and low-visibility scenarios.
Meanwhile, many researchers have explored the use of unpaired image translation techniques to enhance model generalization. By employing unpaired image translation, autonomous vehicles can be trained and tested in a broad range of simulated environments without the need for paired datasets, thus reducing the complexity and cost associated with data collection (Lan et al., 2024; Zhao et al., 2024). Lee et al. (2022) proposed a method that uses image translation to simulate weather effects, such as rain and fog, by transforming LiDAR data between different weather scenarios. This approach facilitates the creation of synthetic data, which augments training datasets and enhances a vehicle’s ability to handle challenging environmental conditions, thereby improving the robustness of autonomous driving systems. Fernando et al. (2024) addressed the challenge of limited visibility in low-light conditions for autonomous driving systems by proposing a CycleGAN-based framework that integrates RGB and infrared (IR) imagery to perform night-to-day image translation. Operating within an unsupervised learning paradigm, the model effectively circumvents the need for labor-intensive paired datasets. Similarly, El Madawi et al. (2019) investigated the fusion of RGB images and LiDAR data for 3D semantic segmentation, leveraging the complementary nature of camera-derived color information and LiDAR-based depth measurements to enhance object recognition accuracy. The aforementioned approaches utilize unpaired image translation techniques for autonomous driving scene generation, introducing IR or LiDAR modalities into a CycleGAN baseline to impose additional structural constraints on the generative adversarial network, thereby improving the realism and fidelity of the synthesized scenes. However, the reliance on auxiliary modalities such as IR or LiDAR inevitably increases data acquisition complexity and cost, which can impose a substantial burden in the context of large-scale scenario generation for autonomous driving system testing.
Therefore, this study adopts an unpaired image translation network as the baseline to enhance the model’s generalization performance. At the same time, it investigates whether high-quality autonomous driving scenario generation can be achieved without incurring additional data acquisition costs. To this end, we propose a dual-branch architecture based on CycleGAN, which integrates monocular depth estimation into the original RGB images, thereby improving both the visual quality and structural integrity of the translated scenes.
3 Main work
In this section, we first describe the two-branch network structure with the original image and its depth map as input, and explain the depth map generation method. We use a dual-branch generator and discriminator, incorporating adversarial loss and cycle consistency loss to constrain the training, thereby achieving style transfer between different datasets. Additionally, we provide details of the generator and discriminator, including the module composition and input–output channel parameters of both networks. Finally, we explain the principle and composition of the added self-attention mechanism.
3.1 Dual-branch network architecture with depth information
In this study, we aim to learn two domain translation functions, referred to as and , which are responsible for mapping an input image , along with its corresponding depth map , to a synthetic image . The objective is to ensure that not only preserves the spatial and structural consistency inherent in , but also effectively adopts the stylistic characteristics of the target domain . The overall framework of the network is shown in Supplementary Figure 1.
As an initial step, we employ the Depth Anything model to estimate depth information from real-world input images, serving as a means to provide geometric guidance for subsequent translation tasks. Depth Anything (Yang S. et al., 2024) trains a robust monocular depth estimation model using large-scale unlabeled datasets, and it has been shown to be state-of-the-art among a large number of monocular depth estimation models.
After obtaining the depth map, we input both the real image and its corresponding depth map into to produce the translated image . To improvice the quality of , we employ , which takes both and as input and computes the adversarial loss .
In addition, we use the pre-trained to reconstruct the input by feeding it and , yielding . The cycle consistency loss is then computed as the expected pixel-wise difference between and . Similarly, we obtain the identity loss by passing and into to generate , and computing the expectation of pixel-wise differences between and .
All loss terms are backpropagated to optimize . The design and training procedure for follow the same principle in the reverse direction.
3.2 Generator and discriminator details
Generator is a dual-branch network composed of two identical networks, the framework of is shown in Supplementary Figure 2A, and the detailed architectures of the branches can be found in Supplementary Table 1. Each branch consists of feature extraction layers, ResNet block layers, self-attention layers, and upsampling layers. The feature extraction layers contain three convolutional layers, which are used to extract features from the input and reduce the image size. The ResNet block layers consist of 9 ResNet blocks, which are used to further extract information from the input features. The self-attention layer introduces a self-attention module to enhance the model’s focus on key information in the input image. Finally, we use upsample layers to increase the image size through the ConvTranspose module and a convolutional module. Additionally, using the residual connection, we add the outputs of the two networks and divide by 2 to obtain the output .
As for the discriminator, its structure is identical to that of the PGGAN discriminator (Karras et al., 2017). This choice is motivated by PatchGAN’s effectiveness in focusing on high-frequency structure, as it operates on local image patches (typically 70 × 70 receptive fields), making it well-suited for capturing texture and style-level details rather than global image semantics. The framework of discriminator is shown in Supplementary Figure 2B, and the detailed architectures of the branches can be found in Supplementary Table 2.
Structurally, the discriminator is implemented as a convolutional neural network composed of multiple layers of strided convolutions, each followed by a LeakyReLU activation function. The number of feature maps increases progressively with network depth, typically doubling at each subsequent layer until reaching a maximum of 512 channels. Instead of producing a single scalar output for the entire input image, the network generates a one-channel feature map, where each spatial location represents the likelihood that the corresponding image patch is real or fake. This patch-based prediction mechanism encourages the generator to produce visually convincing fine-grained details across the entire image, thereby improving the overall realism of the synthesized outputs.
3.3 Implementation details of the depth map module
In the field of unpaired image-to-image translation, to the best of our knowledge, there is currently no method that explicitly utilizes depth map information as input to constrain the translation process. The absence of depth constraints often results in the loss of structural information in the generated images, leading to phenomena such as mode collapse and distortion. Incorporating depth map information can effectively enhance the quality of model training and improve the structural consistency of generated outputs. Therefore, we propose to integrate depth maps as part of the input to our framework.
Our depth estimation method is based on the monocular depth estimation approach proposed in Depth Anything, as illustrated in Supplementary Figure 3. Specifically, we utilize the pretrained model released by Depth Anything to generate depth maps for the input images without making any modifications. The depth map module is designed to leverage both labeled and unlabeled images for training. The labeled dataset and unlabeled dataset are denoted as and , respectively. In the first stage, a teacher model is trained on the labeled dataset using an affine-invariant depth loss to account for variations in scale and shift. The affine-invariant loss between the predicted depth and the ground truth depth is formulated as:
where the predicted and ground truth depths are normalized by subtracting the median and dividing by the mean absolute deviation to eliminate global scale and shift differences.
The trained teacher model is then used to generate pseudo-depth labels for the unlabeled dataset. In the second stage, a student model is trained by combining labeled samples and pseudo-labeled unlabeled samples, enhancing generalization across diverse scenes.
It is important to note that the teacher and student models share an identical network architecture. The encoder is initialized with a pretrained DINOv2 (Oquab et al., 2023) ViT-L model that extracts multi-scale feature representations, and the decoder consists of a DPTHead module that reconstructs dense depth maps from these intermediate features. The overall framework of the model is shown in Supplementary Figure 4. To further improve the quality of the generated depth maps, two types of strong perturbations are introduced during the student training phase: color-based distortions such as color jittering and Gaussian blur, and spatial distortions implemented via CutMix (Yun et al., 2019), which is formulated as:
Where is a binary mask and denotes element-wise multiplication.
Additionally, to enhance semantic understanding and improve feature robustness, a feature alignment loss is applied between the features extracted by the depth encoder and the features extracted by a frozen DINOv2 model. This loss encourages the student model to inherit high-level semantic priors and is defined as:
Where and denote the feature vectors from the student model and the frozen DINOv2 model at pixel , respectively.
Finally, semantic segmentation labels are assigned to the unlabeled images to provide additional supervision. These semantic annotations are generated by combining predictions from RAM (Zhang Y. et al., 2024), GroundingDINO (Liu et al., 2025), and HQ-SAM (Ke et al., 2023), further enhancing the semantic representation learned by the model.
3.4 Self-attention module
Attention mechanisms have been widely applied to various computer vision tasks. Compared to traditional convolutional networks, attention mechanisms are capable of capturing global information from input data, such as extracting contextual relationships in natural language processing tasks, while also focusing on important regions within a broader context (Devlin et al., 2019).
In this paper, we incorporate a self-attention module, with its specific structure illustrated in Supplementary Figure 5. As shown in the figure, the self-attention module consists of two stages, where feature extraction and information fusion are applied to the output of the ResNet block.
In Stage I, the input feature map is transformed through three convolutional layers to generate the query matrix , key matrix , and value matrix . The transformation is formulated as:
In Stage II, similarity matching is performed between the query and key matrices to compute attention weights, which are then fused with the value matrix to obtain the final output of the self-attention module. The matrix operation for Stage II is expressed as follows:
Among them, is the query matrix, is the key matrix, is the value matrix, is the dimension of matrix , the three matrices have variables are divided into the head, the use of linear transformation conversion.
4 Dataset and evaluation
4.1 Experimental data
For this study, we constructed a custom dataset comprising road scene imagery captured using dashcams, as shown in Supplementary Figure 6. The top two rows depict examples from the source domain dataset, which includes elements commonly found in autonomous driving scenarios such as pedestrians, bicycles, delivery vehicles, sedans, and buses. This dataset was collected using a dashcam and captures a variety of road conditions including urban streets and highways. It consists of a total of 6,480 images and forms the original dataset, denoted as .
The middle two rows show sample images from the target domain dataset, which also include various scenes and elements from everyday life. This dataset is derived from the open-source Foggy Cityscapes dataset (Cai et al., 2019) and contains a total of 3,457 foggy images. It is used as the style dataset, denoted as .
In addition, we further utilized the open-source KITTI dataset (Geiger et al., 2012) as another source domain dataset. Corresponding foggy versions of the images were generated to evaluate the generalization ability of our method. The bottom two rows in Supplementary Figure 6 display sample images from the KITTI dataset, which contains a total of 7,071 images.
4.2 Evaluation metrics
4.2.1 Methods for comparison
We compare our method with five existing approaches and conduct a series of experiments, including ablation studies, hyperparameter comparisons, visual comparisons of the generated images, and visual comparisons of the corresponding depth maps. The methods in our comparative evaluation are the follows:
• DCGAN (Radford et al., 2016) stands for Deep Convolutional Generative Adversarial Network, which utilizes convolutional layers to improve the stability and quality of image generation in GANs.
• CoGAN (Liu and Tuzel, 2016) consists of two GANs that share weights to learn a joint distribution of multi-domain images, enabling generation of corresponding images in different domains.
• Stylegan2 (Karras et al., 2020) is a state-of-the-art generative adversarial network for producing highly realistic and high-resolution images through advanced style-based architecture.
• Pix2Pix (Isola et al., 2017) is a conditional generative adversarial network designed for image-to-image translation tasks, capable of transforming images from one domain to another using paired training data.
• Cyclegan (Zhu et al., 2017) is a generative adversarial network designed for unpaired image-to-image translation, allowing conversion between two image domains without needing paired examples.
4.2.2 Metrics
We choose PSNR (Peak Signal-to-Noise Ratio), SSIM (Structural Similarity) and FID (Fréchet Inception Distance) as quantitative indicators.
PSNR is an index used to measure the effect of image compression or denoising, which evaluates the image quality by calculating the mean square error (MSE) between the original image and the compressed or processed image. It is defined as:
Where is the maximum possible pixel value of the image (typically 255), and is the mean squared error between the generated image and the reference image. The larger the PSNR value, the better.
SSIM is used to evaluate the structural similarity of two images. It takes into account the brightness, contrast and structure information of the image, which is closer to the perception of image quality by the human eye. The SSIM is computed as:
Where , are the average values, , are the variances, and is the covariance of images and . and are constants to stabilize the division. The larger the SSIM, the better.
FID is a metric used to evaluate the quality of the generated image and is often used to evaluate the performance of GANs. It measures quality by calculating the difference between the distribution of the generated image and the real image in the feature space. The formula is:
Where () and () are the means and covariances of the real and generated image features, respectively. A smaller FID value indicates a higher similarity to real images and thus better performance.
4.3 Experimental verification
4.3.1 Training strategy and loss function
During the training process, input images are initially resized to a resolution of 286 pixels and subsequently center-cropped to 256 pixels to conform to the input dimensional requirements of the network. To enhance data diversity and improve the model’s generalization capability, particularly with respect to unseen perspectives and spatial configurations, random horizontal flipping is applied by default.
More critically, CycleGAN adopts an asymmetric update strategy to balance the optimization dynamics between the generator and the discriminator. Given that the generator is tasked with learning complex high-dimensional mappings, its training is inherently more challenging, especially during the early phases of learning. In contrast, the discriminator tends to converge more rapidly, which may lead to overly confident predictions and result in vanishing gradients for the generator. To mitigate this imbalance and stabilize the adversarial training process, the update frequency is carefully controlled: the generator is updated three times for every single update of the discriminator.
To ensure both realism and geometric consistency in the unpaired image translation task, particularly under adverse conditions for autonomous driving scenarios, our model employs a composite loss function that integrates adversarial learning, cycle consistency, identity preservation, and depth-aware constraints. These loss components collectively enable robust style transfer while maintaining structural fidelity, which is critical for simulating realistic driving environments.
The total objective function of the proposed Dual-Branch Depth-Aware GAN is expressed as:
Where and , denote the forward and backward generators, respectively. denotes the source domain datasets and denotes the target domain datasets. and are the corresponding discriminators. and are hyperparameters controlling the strength of the cycle consistency and identity losses.
To encourage the generators to produce outputs that are indistinguishable from the real images of the target domain, we adopt the least-squares adversarial loss (LSGAN), which has been shown to stabilize training and generate higher-quality results. For the generator and discriminator , the loss is defined as:
To enforce that the learned mappings are bijective and to preserve semantic structure, we use a cycle consistency loss. This loss ensures that an image translated from one domain to another and then back should closely resemble the original input. It is given by:
To regularize the mapping and maintain color and content consistency, particularly in scenarios where images are already in the target domain, we adopt an identity loss:
This encourages the generators to behave as identity mappings when fed images from the target domain, improving semantic fidelity in the translated outputs.
4.3.2 Comparison against other methods
As shown in Supplementary Table 3, the performance of our proposed method and five comparison methods is evaluated using three quantitative metrics on the custom-designed dataset specifically curated for this study. The results indicate that our approach consistently outperforms the baselines across all evaluation indicators. In particular, the Structural Similarity Index Measure (SSIM) of the images generated by our method reaches 0.80, which is significantly higher than that of the other methods, demonstrating superior image quality and structural preservation.
In addition to the quantitative results, we provide qualitative comparisons of the generated images, as illustrated in Supplementary Figure 7. The leftmost column displays the original input images, the middle columns show the results produced by the baseline methods, and the rightmost column presents the images generated by our proposed method. From the figure, it can be observed that DCGAN and CoGAN struggle to generate coherent and realistic road scene images. While StyleGAN2 and Pix2Pix offer moderate improvements, they often produce mismatches and distortions due to a lack of strong constraints on image structure. CycleGAN shows further enhancement in visual quality, but still suffers from issues such as local artifacts, element distortion, and the unintended addition of scene elements.
In contrast, our method integrates depth map information and a self-attention mechanism into the generation process, which effectively mitigates the above issues. It better preserves semantic elements from the original images, accurately captures scene structure, and significantly reduces artifacts and distortions during translation. These results validate the effectiveness of our design in enhancing both the visual realism and structural consistency of translated images.
4.3.3 Ablation experiment
To evaluate the effectiveness of our framework design, we conduct a comprehensive ablation study using the custom-designed dataset. Specifically, we remove individual components from the proposed method one at a time to assess their respective contributions. The ablated versions of our method are defined as follows:
• m/oself-att: The proposed method without the self-attention component.
• m/odepth: The proposed method without the depth map and dual-branch network structure.
• m/oboth: The proposed method without the depth map, dual-branch network structure and self-attention component.
• m: The proposed method.
The test results, evaluated using PSNR, SSIM, and FID metrics, are presented in Supplementary Figure 8. Several key observations can be drawn from these results:
1. In terms of PSNR and SSIM, our proposed method achieves the best experimental performance. For FID, while our method also shows significant improvement, it ranks second, slightly behind the model incorporating the self-attention mechanism, with a minor gap between the two approaches.
2. Regarding PSNR, the addition of both depth map and self-attention improves the generated image quality, with the depth map having a more pronounced effect. For SSIM, the improvement from both additions is similar. In contrast, the FID score benefits more from the inclusion of the self-attention mechanism.
3. As indicated by the results, incorporating the depth map is particularly effective in image denoising, enhancing the realism and stability of the generated images. Meanwhile, the self-attention mechanism is more effective in capturing the distribution characteristics and structural information of the images.
The experimental results demonstrate that the framework we designed outperforms others across all evaluation metrics. The attention mechanism, known for expanding the receptive field and improving the algorithm’s ability to extract crucial features, contributes significantly to the enhancement of the model’s performance. The dual-branch structure with added depth map information plays a crucial role in providing effective constraints to the generation network, allowing it to produce images consistent with the depth map, which improves the overall quality of the generated output.
To validate our hypothesis, we provide result diagrams and generated images comparing the proposed method with the dual-branch network method without depth map information, as shown in Supplementary Figure 9. From the figure, it is evident that our proposed method preserves the road elements well, while the method without depth maps fails to maintain background road features effectively. Moreover, the latter may introduce elements not present in the original road scene, such as the addition of a car logo in the original image , or the appearance of a car in the middle of the road in image . Additionally, the vehicle shape in the original image becomes distorted.
From the experimental results, it is clear that the dual-branch network incorporating depth map information better maintains the features of road elements and effectively prevents issues such as distortion, false shadows, and other artifacts in the generated images.
4.3.4 Hyperparameter comparison
We have compared different hyperparameters of the network, and the experimental results are shown in Supplementary Table 4. n_epochs represents the number of iterations in which the learning rate does not decay at the beginning of the training process, and n_epochs_decay represents the number of iterations in which the learning rate decays after the training epochs round. It can be seen from the table that applying learning rate decay significantly affects model training performance. In this problem, n_epochs = 50, n_epochs_decay = 50, the generation effect is the best. lambdaA and lambdaB represent the weights of the generator and discriminator during training, respectively, and the best effect is achieved when lambdaA = 100 and lambda = 100. Identity indicates the weights of Indentity_loss and reconstruction loss in the training process. As can be seen from the table, the best result is generated when Identity = 0.5.
4.3.5 Presentation of results
Supplementary Figure 10 illustrates the results of road scene generation using the proposed model using the custom-designed dataset. The first row displays the original images; the second row shows the corresponding depth maps generated by the Depth Anything model; the third row presents samples from the style dataset , selected from the Foggy Cityscapes dataset; and the final row shows the foggy images generated by our method.
As observed from the figure, the images produced by our model not only effectively preserve the semantic elements and structural integrity of the original road scenes. Such as road markings, vehicles, and surrounding objects, but also successfully learn and replicate the foggy style characteristics of the target domain . This demonstrates the model’s strong ability in both content preservation and style transfer.
To ensure the objectivity of the evaluation, the same set of source images was used across all methods, and visual comparisons were supplemented with quantitative metrics as presented in previous sections. The consistency between the learned depth structure and the transferred style highlights the advantage of integrating depth information in the generative process.
In addition, we further employed the open-source KITTI dataset as the source domain dataset to generate foggy-style images using the pretrained weights obtained from our model trained on the custom dashcam dataset. The generation results are presented in Supplementary Figure 11. As shown in the figure, the proposed method successfully applies fog-style effects to the original scenes while preserving the underlying structural content. These results not only demonstrate the effectiveness of the style transfer but also validate the generalization capability of the model across different datasets.
5 Conclusion
Reliable visual perception is essential for autonomous driving, as accurate scene understanding is crucial for safe decision-making in dynamic and challenging environments. However, adverse weather conditions such as fog, rain, and low-light environments introduce significant challenges to the robustness and generalization of perception models. Traditional unpaired image-to-image translation methods, which primarily rely on RGB-based transformations, often lead to geometric distortions and loss of structural consistency. These issues are especially problematic in safety-critical applications such as autonomous driving. To address these challenges, we propose the Depth-Aware Dual-Branch Generative Adversarial Network (DAB-GAN), which integrates depth information directly into the generative process. This approach preserves spatial structures and enhances translation accuracy in adverse conditions.
Our DAB-GAN model utilizes a dual-branch generator that processes both RGB and depth inputs, ensuring geometric fidelity while leveraging a self-attention mechanism to enhance spatial dependencies and refine local details. This integration allows for the generation of more realistic and structurally coherent images that are vital for autonomous driving perception systems. The incorporation of depth-aware structural integrity significantly improves the generalization capabilities of vision-based models, particularly in dynamic environments and challenging weather conditions.
Through extensive experiments, we demonstrate that DAB-GAN outperforms existing unpaired image-to-image translation methods across multiple benchmarks, achieving superior visual fidelity and structural consistency. Our results validate the effectiveness of combining depth-awareness and self-attention mechanisms to improve the quality and realism of generated images, ensuring that the translated images remain both visually realistic and geometrically faithful.
This work highlights the potential of depth-aware generative models to address real-world challenges in autonomous driving systems. Future work will explore optimization techniques for real-time applications, as well as the fusion of multi-sensor data to further enhance translation accuracy and system robustness in real-world autonomous driving deployments.
Data availability statement
Publicly available datasets were analyzed in this study. This data can be found at: MTgzMjMxMzU0NzFAMTYzLmNvbQ== to get the data.
Author contributions
DS: Conceptualization, Formal analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft. CXZ: Conceptualization, Funding acquisition, Methodology, Supervision, Writing – review & editing. CBZ: Funding acquisition, Investigation, Methodology, Resources, Supervision, Writing – review & editing. ZF: Conceptualization, Investigation, Visualization, Writing – original draft. CY: Conceptualization, Investigation, Methodology, Writing – review & editing. JL: Conceptualization, Funding acquisition, Resources, Visualization, Writing – review & editing. MF: Conceptualization, Data curation, Supervision, Validation, Visualization, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This research was funded by the National Key R&D Program of China (grant No.2022YFF0604803), the Key R&D Program of Zhejiang Province, China (grant No.2022C01050, 2023C01238), the science and technology project of Zhejiang Province Market Supervision Administration, China (grant No. QN2023426, CY2022339, CY202310, ZC2023001, ZD2024010), and the Basic Public Welfare Research Program of Zhejiang Province, China (grant NO. LGC22E050004).
Acknowledgments
We extend our deepest gratitude to the following researchers, listed alphabetically by first name: Daofei Li, Weiwen Deng, and other contributing researchers. Their contributions have greatly advanced and promoted research in this field.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnbot.2025.1603964/full#supplementary-material
References
Agarwal, S., Birman, R., and Hadar, O. (2025). “WarLearn: weather-adaptive representation learning,” in 2025 IEEE Winter Conference on Applications of Computer Vision, 4978–4987. doi: 10.1109/WACV61041.2025.00487
Bakht, A., Jia, Z., Din, M., Akram, W., Saoud, L., Seneviratne, L., et al. (2024). Mula-Gan: multi-level attention Gan for enhanced underwater visibility. Ecol. Inform. 81:102631. doi: 10.1016/j.ecoinf.2024.102631
Biagiola, M., and Tonella, P. (2024). Boundary state generation for testing and improvement of autonomous driving systems. IEEE Trans. Softw. Eng. 50, 2040–2053. doi: 10.1109/TSE.2024.3420816
Cai, Q., Pan, Y., Ngo, C., Tian, X., Duan, L., and Yao, T. (2019). “Exploring object relation in mean teacher for cross-domain detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 11449–11458. doi: 10.1109/CVPR.2019.01172
Cai, J., Yang, S., and Guang, H. (2024). A review on scenario generation for testing autonomous vehicles. IEEE Intell. Veh. Symp., 3371–3376. doi: 10.1109/IV55156.2024.10588675
Cai, X., Zhu, Y., Miao, D., Fu, L., and Yao, Y. (2024). “Rethinking the paradigm of content constraints in unpaired image-to-image translation,” in Proceedings of the AAAI Conference on Artificial Intelligence, 891–899. doi: 10.1609/aaai.v38i2.27848
Chen, L., Jing, Q., Zhou, Y., Li, Z., Shi, L., and Sun, L. (2024). Element-conditioned Gan for graphic layout generation. Neurocomputing 591:127730. doi: 10.1016/j.neucom.2024.127730
Chen, M., Zheng, Z., and Yang, Y. (2024). “Transferring to real-world layouts: a depth-aware framework for scene adaptation,” in Proceedings of the 32nd ACM International Conference on Multimedia, 399–408. doi: 10.1145/3664647.3681041
Devlin, J., Chang, M., Lee, K., and Toutanova, K. (2019). “Bert: pre-training of deep bidirectional transformers for language understanding,” in Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 1, 4171–4186. doi: 10.18653/v1/N19-1423
Ding, J., Peng, Y., Huang, M., and Zhou, S. (2024). Agrigan: unpaired image dehazing via a cycle-consistent generative adversarial network for the agricultural plant phenotype. Sci. Rep. 14:14994. doi: 10.1038/s41598-024-65540-0
El Madawi, K., Rashed, H., El Sallab, A., Nasr, O., Kamel, H., and Yogamani, S. (2019). “Rgb and lidar fusion based 3d semantic segmentation for autonomous driving,” in 2019 IEEE intelligent transportation systems conference, 7–12. doi: 10.1109/ITSC.2019.8917447
Fernando, K., Lakmal, H., Dissanayake, M., and Kalansooriya, L. (2024). “Enhancing perception for autonomous driving systems with dual-vision input cyclegan for night-to-day image translation,” in 8th SLAAI International Conference on Artificial Intelligence, 1–6. doi: 10.1109/SLAAI-ICAI63667.2024.10844932
Geiger, A., Lenz, P., and Urtasun, R. (2012). “Are we ready for autonomous driving? The Kitti vision benchmark suite,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 3354–3361. doi: 10.1109/CVPR.2012.6248074
Goodfellow, I., Pouget-abadie, J., Mirza, M., Xu, B., Warde-farley, D., Ozair, S., et al. (2014). Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27, 2672–2680. doi: 10.48550/arXiv.1406.2661
Hong, F., Shen, L., and Xu, D. (2024). Dagan plus plus: depth-aware generative adversarial network for talking head video generation. IEEE Trans. Pattern Anal. Mach. Intell. 46, 2997–3012. doi: 10.1109/TPAMI.2023.3339964
Hong, F., Zhang, L., Shen, L., and Xu, D. (2022). “Depth-aware generative adversarial network for talking head video generation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 3387–3396. doi: 10.1109/CVPR52688.2022.00339
Huang, Y., Sun, J., and Tian, Y. (2025). A bayesian optimization method for finding the worst-case scenarios of autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 26, 529–543. doi: 10.1109/TITS.2024.3490616
Ioannou, E., and Maddock, S. (2022) “Depth-aware neural style transfer using instance normalization,” in Proceedings of the Computer Graphics and Visual Computing, 1–8. doi: 10.2312/cgvc.20221165
Isola, P., Zhu, J., Zhou, T., and Efros, A. (2017). “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5967–5976. doi: 10.1109/cvpr.2017.632
Karras, T., Aila, T., and Laine, S., (2017). “Progressive growing of gans for improved quality, stability, and variation,” in Proceedings of the International Conference on Learning Representations. doi: 10.48550/arXiv.1710.10196
Karras, T., Aittala, M., Laine, S., Härkönen, E., Hellsten, J., Lehtinen, J., et al. (2021a). Alias-free generative adversarial networks. Adv. Neural Inf. Process. Syst. 34, 852–863. doi: 10.48550/arXiv.2106.12423
Karras, T., Laine, S., and Aila, T. (2021b). A style-based generator architecture for generative adversarial networks. IEEE Trans. Pattern Anal. Mach. Intell. 43, 4217–4228. doi: 10.1109/TPAMI.2020.2970919
Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., and Aila, T. (2020). “Analyzing and improving the image quality of stylegan,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 8107–8116. doi: 10.1109/CVPR42600.2020.00813
Ke, L., Ye, M., Danelljan, M., Liu, Y., Tai, Y., Tang, C., et al. (2023). Segment anything in high quality. Adv. Neural Inf. Process. Syst. 36, 29914–29934. doi: 10.48550/arXiv.2306.01567
Lakmal, H., Dissanayake, M., and Aramvith, S. (2024). Light the way: an enhanced generative adversarial network framework for night-to-day image translation with improved quality. IEEE Access 12, 165963–165978. doi: 10.1109/access.2024.3491792
Lambertenghi, S., and Stocco, A. (2024). “Assessing quality metrics for neural reality gap input mitigation in autonomous driving testing,” in Proceedings of the IEEE International Conference on Software Testing, Verification, and Validation, 173–184. doi: 10.1109/ICST60714.2024.00024
Lan, G., Peng, Y., Hao, Q., and Xu, C. (2024). Sustechgan: image generation for object detection in adverse conditions of autonomous driving. IEEE Trans. Intell. Veh., 1–10. doi: 10.1109/TIV.2024.3504566
Lee, J., Shiotsuka, D., Nishimori, T., Nakao, K., and Kamijo, S. (2022). Gan-based lidar translation between sunny and adverse weather for autonomous driving and driving simulation. SENSORS-BASEL 22:5287. doi: 10.3390/s22145287
Li, A., Chen, S., Sun, L., Zheng, N., Tomizuka, M., and Zhan, W. (2022). Scegene: bio-inspired traffic scenario generation for autonomous driving testing. IEEE Trans. Intell. Transp. Syst. 23, 14859–14874. doi: 10.1109/TITS.2021.3134661
Li, Y., Lin, Z., Forsyth, D., Huang, J., and Wang, S. L. (2023). “Climatenerf: extreme weather synthesis in neural radiance field,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 3204–3215. doi: 10.1109/ICCV51070.2023.00299
Li, X., Teng, S., Liu, B., Dai, X., Na, X., and Wang, F. (2023). Advanced scenario generation for calibration and verification of autonomous vehicles. IEEE Trans. Intell. Veh. 8, 3211–3216. doi: 10.1109/TIV.2023.3269428
Lin, Q., Li, Z., Zeng, K., Wen, J., Jiang, Y., and Chen, J. (2025). Wtngan: unpaired image translation from white light images to narrow-band images. Pattern Recogn. 162:111431. doi: 10.1016/j.patcog.2025.111431
Liu, M., Breuel, T., and Kautz, J. (2017). Unsupervised image-to-image translation networks. Adv. Neural Inf. Process. Syst. 30, 700–708. doi: 10.48550/arXiv.1703.00848
Liu, X., Cheng, M., Lai, Y., and Rosin, P. (2017) “Depth-aware neural style transfer,” in Proceedings of Symposium on Non-Photorealistic Animation and Rendering, 4, 1–10. doi: 10.1145/3092919.3092924
Liu, Y., Liao, S., Zhu, Y., Deng, F., Zhang, Z., Gao, X., et al. (2024). Channel-spatial attention guided cyclegan for cbct-based synthetic ct generation to enable adaptive radiotherapy. IEEE Trans. Comput. Imaging 10, 818–831. doi: 10.1109/TCI.2024.3402372
Liu, M., and Tuzel, O. (2016). Coupled generative adversarial networks. Adv. Neural Inf. Process. Syst. 29, 469–477. doi: 10.48550/arXiv.1606.07536
Liu, S., Zeng, Z., Ren, T., Li, F., Zhang, H., Yang, J., et al. (2025). “Grounding dino: marrying dino with grounded pre-training for open-set object detection,” in Proceedings of the European Conference on Computer Vision, 15105, 38–55. doi: 10.1007/978-3-031-72970-6_3
Ming, Y., Meng, X., Fan, C., and Yu, H. (2021). Deep learning for monocular depth estimation: a review. Neurocomputing 438, 14–33. doi: 10.1016/j.neucom.2020.12.089
Oquab, M., Darcet, T., Moutakanni, T., Vo, H., Szafraniec, M., Khalidov, V., et al. (2023). Dinov2: learning robust visual features without supervision. arxiv [Preprint]. arxiv:2304.07193. doi: 10.48550/arXiv.2304.07193
Piccinelli, L., Yang, Y., Sakaridis, C., Segu, M., Li, S., Van, G., et al. (2024). “Unidepth: universal monocular metric depth estimation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 10106–10116. doi: 10.1109/CVPR52733.2024.00963
Radford, A., Metz, L., and Chintala, S. (2016). Unsupervised representation learning with deep convolutional generative adversarial networks. arxiv [Preprint]. arxiv:1511.06434. doi: 10.48550/arXiv.1511.06434
Sadid, H., and Antoniou, C. (2024). A simulation-based impact assessment of autonomous vehicles in urban networks. IET Intell. Transp. Syst. 18, 1677–1696. doi: 10.1049/itr2.12537
Shrivastava, A., Pfister, T., Tuzel, O., Susskind, J., Wang, W., and Webb, R. (2017). “Learning from simulated and unsupervised images through adversarial training,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2242–2251. doi: 10.1109/CVPR.2017.241
Shubham, K., Sastry, P., and Prathosh, A. (2024). “Fusing conditional submodular Gan and programmatic weak supervision,” in Proceedings of the AAAI Conference on Artificial Intelligence, 15020–15028. doi: 10.1609/aaai.v38i13.29423, 38
Tong, W., Cai, Y., Jie, Y., Duan, Y., Hou, Y., Wu, E., et al. (2025a). Neural rendering and flow-assisted unsupervised multi-view stereo for real-time monocular tracking and scene perception. IEEE Trans. Autom. Sci. Eng. 1:1. doi: 10.1109/tase.2025.3546713
Tong, W., Guan, X., Zhang, M., Li, P., Ma, J., Wu, E., et al. (2025b). Edge-assisted epipolar transformer for industrial scene reconstruction. IEEE Trans. Autom. Sci. Eng. 22, 701–711. doi: 10.1109/tase.2023.3330704
Tong, W., Zhang, M., Zhu, G., Xu, X., and Wu, E. (2024). Robust depth estimation based on parallax attention for aerial scene perception. IEEE Trans. Ind. Inform. 20, 10761–10769. doi: 10.1109/tii.2024.3392270
Wei, H., Wu, Q., Wu, C., Ngan, K., Li, H., Meng, F., et al. (2024). Robust unpaired image dehazing via adversarial deformation constraint. IEEE Trans Circ. Syst. Vid. 34, 8614–8628. doi: 10.1109/TCSVT.2024.3387451
Xue, Y., Chen, K., and Neri, F. (2024). Differentiable architecture search with attention mechanisms for generative adversarial networks. IEEE Trans Emerg. Top. Comput. Intell. 8, 3141–3151. doi: 10.1109/TETCI.2024.3369998
Yang, L., Kang, B., Huang, Z., Xu, X., Feng, J., and Zhao, H. (2024). “Depth anything: unleashing the power of large-scale unlabeled data,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 10371–10381. doi: 10.1109/CVPR52733.2024.00987
Yang, S., Qin, H., Yuan, S., Yan, X., and Rahmani, H. (2024). Destripecyclegan: stripe simulation cyclegan for unsupervised infrared image destriping. IEEE Trans Instrum and Meas. 73, 1–14. doi: 10.1109/TIM.2024.3476560
Ye, R., Boukerche, A., Yu, X., Zhang, C., Yan, B., and Zhou, X. (2024). Data augmentation method for insulators based on cycle-Gan. J. Electron. Sci. Technol. 22:100250. doi: 10.1016/j.jnlest.2024.100250
Ye, M., Meng, Z., and Qian, Y. (2024). Building cross-domain mapping chains from multi-cyclegan for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 62, 1–17. doi: 10.1109/TGRS.2024.3431460
Ye, X., and Wang, K. (2024). Deep generative domain adaptation with temporal relation attention mechanism for cross-user activity recognition. Pattern Recogn. 156:110811. doi: 10.1016/j.patcog.2024.110811
Ye, H., Xiang, H., and Xu, F. (2024). Cycle-Gan network incorporated with atmospheric scattering model for dust removal of martian optical images. IEEE Trans. Geosci. Remote Sens. 62, 1–13. doi: 10.1109/TGRS.2024.3432601
Yildiz, E., Yuksel, M., and Sevgen, S. (2024). A single-image Gan model using self-attention mechanism and densenets. Neurocomputing 596:127873. doi: 10.1016/j.neucom.2024.127873
Yun, S., Han, D., Oh, S., Chun, S., Choe, J., and Yoo, Y. (2019). “Cutmix: regularization strategy to train strong classifiers with localizable features,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 6022–6031. doi: 10.1109/ICCV.2019.00612
Zhang, Y., Huang, X., Ma, J., Li, Z., Luo, Z., Xie, Y., et al. (2024). “Recognize anything: a strong image tagging model,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1724–1732. doi: 10.1109/cvprw63382.2024.00179
Zhang, J. W., Xu, C. J., and Li, B. (2024). “Chatscene: knowledge-enabled safety-critical scenario generation for autonomous vehicles,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 15459–15469. doi: 10.1109/CVPR52733.2024.01464
Zhao, H., Wang, Y., Bashford-rogers, T., Donzella, V., and Debattista, K. (2024). Exploring generative ai for sim2real in driving data synthesis. IEEE Intell. Veh. Symp., 3071–3077. doi: 10.1109/IV55156.2024.10588493
Keywords: autonomous driving, unpaired image-to-image translation, depth map, self-attention mechanism, generative adversarial network
Citation: Shi D, Zhao C, Zhao C, Fang Z, Yu C, Li J and Feng M (2025) Depth-aware unpaired image-to-image translation for autonomous driving test scenario generation using a dual-branch GAN. Front. Neurorobot. 19:1603964. doi: 10.3389/fnbot.2025.1603964
Edited by:
Hu Cao, Technical University of Munich, GermanyReviewed by:
Maheshi Dissanayake, University of Peradeniya, Sri LankaWei Tong, Nanjing University of Posts and Telecommunications, China
Naoual Mebtouche, University of Science and Technology Houari Boumediene, Algeria
Wenlong Lu, Shanghai Jiao Tong University, China
Copyright © 2025 Shi, Zhao, Zhao, Fang, Yu, Li and Feng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Chenxin Zhao, emN4XzIwMDlAZm94bWFpbC5jb20=; Cunbin Zhao emhhb2N1bmJpbkAxMjYuY29t