- 1School of Computer Science, Wuhan University, Wuhan, China
- 2Intelligent Transport Systems Research Center, Wuhan University of Technology, Wuhan, China
- 3GNSS Research Center, Wuhan University, Wuhan, China
Although deep learning methods have made significant strides in single image sand dust removal, the heterogeneous uncertainty induced by dusty environments poses a considerable challenge. In response, our research presents a novel framework known as the Hierarchical Interactive Uncertainty-aware Network (HIUNet). HIUNet leverages Bayesian neural networks for the extraction of robust shallow features, bolstered by pre-trained encoders for feature extraction and the agility of lightweight decoders for preliminary image reconstitution. Subsequently, a feature frequency selection mechanism is activated to enhance overall performance by strategically identifying and retaining valuable features while effectively suppressing redundant and irrelevant ones. Following this, a feature enhancement module is applied to the preliminary restoration. This intricate fusion culminates in the production of a restored image of superior quality. Our extensive experiments, using our proposed Sand11K dataset that exhibits various levels of degradation from dust and sand, confirm the effectiveness and soundness of our proposed method. By modeling uncertainty via Bayesian neural networks to extract robust shallow features and selecting valuable features through frequency selection, HIUNet can reconstruct high-quality clean images. For future work, we plan to extend our uncertainty-aware framework to handle extreme sand scenarios.
1 Introduction
Adverse weather conditions, including rain, haze, dust, and sandstorms (Shi et al., 2023; Narasimhan and Nayar, 2002; Li et al., 2017b; Shi et al., 2019; Zheng et al., 2021; Zamir et al., 2021, 2022; Wu et al., 2021; Li et al., 2022), present substantial challenges for visual perception systems. Sand and dust are particularly problematic: they obscure critical details like object edges and textures, and directly impair downstream tasks such as object detection and autonomous navigation. This drives significant advancements in image restoration research (Zamir et al., 2021). While deep learning methods, particularly those based on encoder-decoder (Chen et al., 2022) and UNet architectures (Zamir et al., 2022), have demonstrated remarkable progress, they face inherent limitations when handling the complex uncertainty patterns characteristic of sand and dust degradation. Traditional encoder-decoder frameworks (Zheng et al., 2021; Qin et al., 2020; Wu et al., 2021; Zamir et al., 2022), despite their success in general image restoration tasks, exhibit critical shortcomings in extracting comprehensive feature representations from severely degraded images (Zamir et al., 2021), primarily due to their single-level feature extraction paradigm and inadequate uncertainty modeling capabilities.
To address these challenges, we propose a novel framework that integrates uncertainty estimation with shallow feature extraction in a unified, uncertainty-aware architecture. Our approach leverages three key components: (1) Bayesian Neural Networks (BNNs) (Kononenko, 1989; Mackay, 1992; Neal, 2012) for uncertainty-aware shallow feature extraction, (2) a pre-trained encoder (Chen et al., 2022) for robust deep feature extraction, and (3) lightweight decoder modules for initial restoration. The framework further incorporates a frequency-aware feature selection network to dynamically identify and enhance informative features. Following initial restoration, we introduce a shallow feature refinement module that progressively enhances feature representations through thinning operations. Finally, a multi-level feature fusion mechanism synthesizes hierarchical information to generate the final restored image. Extensive experiments on diverse degradation scenarios, including sand, sandstorm, and dust conditions, demonstrate the effectiveness and theoretical soundness of our proposed method.
In summary, the primary contributions of this paper can be categorized into these main aspects:
• To the best of our knowledge, we are the first to utilize uncertainty estimation to characterize various degradation phenomena in single image sand removal, employing the Bayesian neural networks as its implementation. We also first to employ a pre-trained encoder to get deep prior from valuable deep features in single image sand dust removal.
• We introduce a novel frequency selection mechanism for facilitating channel dimensional interaction.
• We propose the Hierarchical Interactive Uncertainty-aware Network(HIUNet) for accurate single image sand dust removal.
• Through comprehensive experiments conducted on our Sand11K dataset, we demonstrate that our proposed method outperforms current state-of-the-art techniques.
In the following section, we will provide the related works of single image sand dust removal.
2 Related works
2.1 Sand images formulation
The extensively applied physical model elucidating the formation of images perturbed by light transmission through haze (Narasimhan and Nayar, 2002; Si et al., 2022) is conventionally delineated as follows:
where I(p) represents the observed hazy image, t(p) signifies the medium transmission map, A denotes the global atmosphere light, and J(p) corresponds to the haze-free image. Further, t(p) can be expressed as: t(p) = e−βd(p), where β represents the scattering coefficient of the atmosphere, and d(p) signifies the scene depth between the digital camera and the captured object for each pixel p in the image.
In dusty conditions, the different ways that red, green, and blue channels degrade lead to telltale signs like shifts, density changes, and variations over time in the images. It can be formulated as:
where AG = k1AR+b1, AB = k2AR+b2, Â is the global color deviation value of the sand dust image, b is the disturbance amount, and k is the spatial distribution coefficient of the atmospheric light value of the three basic color spectrums. Based on Equation 2, various sandy images with different degrees of degradation can be synthesized. The detailed formulation process is shown in Figure 1.

Figure 1. The formulation process of sand and dust images. (a) Dust image, (b) Sand image, (c) Sandstorm image.
2.2 Image recovery from degraded weather
Qin et al. (2020) introduce FFA-Net, a feature fusion attention network that significantly advances single image dehazing by effectively integrating channel and pixel attention mechanisms; Zamir et al. (2022) introduce Restormer, an efficient Transformer equipped with a novel design that mitigates the computational complexity of traditional Transformers. Gao et al. (2024) propose a single-stage design, based on a simple UNet architecture, with a mountain-shaped structure. Chen et al. (2022) present NAFNet, an innovative approach that simplifies image restoration by eliminating nonlinear activations, achieving comparable state-of-the-art results with reduced computational complexity.
2.3 Uncertainty-aware probabilistic modeling
Data uncertainty in deep learning is a key factor influencing model performance, and probabilistic modeling approaches to address data uncertainty are gaining growing attention. Chang et al. (2020) introduce data uncertainty learning in face recognition, proposing DULcls and DULrgs methods that simultaneously optimize for feature extraction and uncertainty. Guo et al. (2022) present the Uncertainty-Guided Probabilistic Transformer (UGPT), introducing a novel probabilistic approach to enhance training and inference under uncertainty in action recognition tasks. Wang et al. (2021) introduce a data-uncertainty guided multi-phase learning approach for semi-supervised object detection that effectively leverages unlabeled data across varying difficulty levels. Yang et al. (2021) present UGTR, a pioneering approach that integrates Bayesian learning with Transformer-based Vaswani (2017) reasoning to enhance the model's performance for camouflaged object detection.
2.4 Bayesian deep learning
In a conventional neural network, the weights are fixed, which leads to its inability to handle the uncertainty in training data effectively. Therefore, regular neural networks cannot handle the uncertainty in training data well. Unlike the weights in traditional neural network models, the weight parameters in Bayesian neural network (Kononenko, 1989) models are no longer a single deterministic value but rather a probability distribution. The learning process of a neural network can be regarded as a Maximum Likelihood Estimation (MLE) (Mackay, 1992):
where D corresponds to the data used for training, xi represents the input of each node, w represents the weights of each node, and yi represents the output. And if a prior is introduced for w, it becomes a Maximum a Posterior(MAP) (Neal, 2012):
Unlike MAP estimation which only provides the optimal solution through argmax computation, Bayesian estimation derives the complete posterior distribution P(w|D) for parameters w. In Bayesian neural networks, weights are characterized as probability distributions rather than deterministic values. This probabilistic formulation enables explicit modeling of prediction uncertainty, which is particularly beneficial for handling diverse degradation patterns in image restoration tasks by quantifying and propagating data uncertainty through the network.
3 Method
3.1 Overall pipeline
We propose HIUNet, a novel framework for removing dust, sand and sandstorm artifacts from single images (Figure 2). Our method contains five key components: First, to model uncertainty probabilistically, we integrate Bayesian neural networks (Wang and Yeung, 2020; Korb and Nicholson, 2010) that quantify degradation-type uncertainties in the input data. Second, an encoder-decoder architecture extracts deep features, where the encoder utilizes pre-trained weights to capitalize on prior visual knowledge. Third, a Feature Selection Network (FSN) dynamically integrates features from dual extractors through value-based selection, emphasizing informative channels while suppressing redundant ones. Fourth, a Cross-scale Feature Enhancement Module (CFEM) progressively refines the coarse outputs from the deep feature extractor through multi-scale contextual fusion. Finally, the enhanced features from CFEM are concatenated with FSN-selected features to produce the final reconstruction. The comprehensive architecture is visualized in Figure 2. Furthermore, the pseudo-code of our HIUNet can be seen in Algorithm 1.

Figure 2. Overall structure of the proposed method. Sharing images are fed into two branches. One branch goes through the shallow feature extractor to generate dynamic weights and then enters the feature selection network with multiple efficient feature selectors. The other branch passes through the deep feature extractor which consists of an encoder and a decoder. After that, the cross-scale feature enhancement module processes these features. The output of the deep feature extractor is also input into the feature selective network. Moreover, the intermediate features of the deep feature extractor are fused into the feature selection network through feature fusion. Finally, the outputs of the two branches are concatenated and then integrated by a convolution layer. Finally, a residual connection is established with the input degraded image to refine the feature integration, culminating in the generation of the final output.
3.2 Uncertainty-aware probabilistic shallow feature extractor
In our proposed method, Bayesian neural networks (Wang and Yeung, 2020; Korb and Nicholson, 2010) are used as an uncertainty-aware probabilistic shallow feature extractor. Due to the task of simultaneously enhancing the quality of a variety of severe weather-degraded images, the data itself have considerable uncertainty. In order to cope with this situation, we use an uncertainty-aware shallow feature extraction module to solve it. This network is designed to extract features corresponding to multiple types of degradation, providing a more nuanced representation of the input data. For input x∈Rh×w×c, the whole process is shown in Equation 5:
where x and are the input and output feature maps, B1 and B2 are the BNN layers in our Uncertainty-aware Probabilistic Shallow Feature Extractor (UPSFE), θ is the Gaussian Error Linear Unit. f3 denotes the convolutional layer with a kernel size 3. The structure of our proposed UPSFE is shown in Figure 3a.

Figure 3. (a) Structure of the proposed Uncertainty-aware Probabilistic Shallow Feature Extractor (UPSFE). (b) Cross-scale Feature Enhancement Module (CFEM). (c) Structure of the employed encoder-decoder.
3.3 Feature selection network
We introduce the Feature Selection Network(FSN), a module dedicated to the preservation of fine details throughout the image restoration process. Engineered to seamlessly transfer intricate details from the degraded input to the output, FSN enhances the restored image with spatial precision and rich contextual coherence. Furthermore, to sift through complex features and distill those of higher value while suppressing the relatively useless ones, we propose a novel Efficient Frequency Selector (EFS). EFS employs a scale-interacting approach to efficiently integrate features. The following formula mathematically represents the feature selection process:
where DS and SG are depth separable convolution and simple gate, σ and ψ are Sigmoid and ReLU activation functions. The convolutional layer with a kernel size k is denoted asfk. [·, ·]denotes the concatenation operation. The detailed framework of our proposed Frequency Selector is shown in Figure 4.

Figure 4. Detailed framework of the feature selection network equipped with our proposed efficient frequency selector.
3.4 Cross-scale feature enhancement module
In our proposed method, we also utilize the output of deep feature extraction to calculate the loss for the entire network. To refine this output, we introduce a cross-scale feature enhancement module. Firstly, we perform convolution operations on features of multiple scales, enabling them to be concatenated. Next, we apply a dimension reduction operation, followed by integrating the channel dimension characteristics through convolution operations on the large-scale channel dimension. Finally, we stitch together the coarse features after concatenation with the refined channel dimension features and assign elements one by one using pixel-scale convolution. The following formula can represent the entire operation:
where is a convolution layer with n kernel and c channel, and the structure of this part is shown in Figure 3b.
3.5 Pre-trained prior deep feature extractor
In image restoration, the UNet architecture is widely utilized. The encoder serves as a feature extractor, while the decoder is responsible for restoring the image resolution. However, most existing UNet-based methods train the entire Encoder-Decoder together, without separately processing the encoder and decoder. We propose a novel method that leverages a pre-trained encoder for feature extraction for deep prior inspired by previous work (He et al., 2022). The pre-trained encoder is pre-trained on our proposed Sand11K dataset for 200 epoches. We split the UNet into two parts, loading only the encoder's weight. Additionally, we have eliminated some feature processing blocks in the middle of the Encoder-Decoder (Chen et al., 2022), as we believe it to be redundant. Moreover, we have improved the decoder to improve its sensitivity to the texture of features by adding Spatial Attention (Woo et al., 2018) as texture perception mechanisms. The formulation of this whole module is expressed through the following formula:
where Enc, Mid and Dec are Encoder, Middle Blocks and Decoder, respectively, and the structure of Encoder-Decoder employed in this paper is shown in Figure 3c.
3.6 Loss function
In order to train all the neural networks involved in this paper in a simple and fair way, we use the Charbonnier penalty function (Charbonnier et al., 1994) to train all the neural networks, including all the comparison experiments and ablation studies. Charbonnier penalty function (Charbonnier et al., 1994) is more tolerant of small errors and holds better convergence during training (Jiang et al., 2020). Charbonnier penalty function (Charbonnier et al., 1994) is shown in:
where I′ is the restored image, is the ground-truth image, and ϵ = 10−3 is a constant in all the experiments.
4 Experiments
4.1 Implementation details
We implement our framework and other State-Of-The-Art (SOTA) methods using PyTorch with a NVIDIA RTX4090 GPU. We train all the models for 200 epochs with AdamW optimizer. The initial learning rate is set to 1 × 10−4, and we employ CosineAnnealingLR to adjust the learning rate. The size of the image patches used for training is 256 × 256. For our proposed Sand11K dataset, following the previous work (Shi et al., 2023), we also synthesize sand and dust images for training, we use 11,985 images as input and 799 ground truth images as labels. The clear images are from PASCAL VOC dataset (Everingham et al., 2015). For testing the performance of learning-based methods, we use 2,397 synthesized sand dust images and real sand dust images.
For the construction process of our Sand11K dataset, we chose the Pascal VOC (Everingham et al., 2015) dataset as clear images. For the training set, we picked 799 sharp images from the Pascal VOC dataset. Then, according to different Atmospheric Light, we synthesized three types of degraded images: dust, sand, and sandstorm, and each type of image contained five degraded images with different degrees according to the Scattering Coefficient. As a result, 1,1985 degraded images were obtained. Then, for the test set, we also selected 799 clear images from the Pascal VOC dataset, and then, according to different Atmospheric Light, we synthesized three types of degraded images: Dust, Sand and Sandstorm. Therefore, the test data totals 2,397 image pairs.
4.2 Experimental results
In the comparative experiments with other state-of-the-art methods, we calculate the scores of Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Natural Image Quality Evaluator (NIQE) (Mittal et al., 2012b) and Blind/Referenceless Image Spatial Quality Evaluator (BRISQUE) (Mittal et al., 2012a) using the RGB channelfor the sand dust removal task of a single image. As shown in Table 1, our method achieves significant performance gains that are consistent with or even more than the SOTA methods. For example, our method outperforms the second-best method M3SNet (Gao et al., 2024) 0.27 dB in PSNR for synthetic image sand dust removel. For real-world image sand dust removal, our mehtod ourperforms the second-best method MPRNet 1.81 in BRISQUE. Meanwhile, we provide the computing complexity of all the methods in this paper as shown in Table 2. Furthermore, the visual results shown in Figure 5 are closely aligned with the quantitative findings, substantiating our proposed method's superior image restoration capabilities. As depicted in Figure 6, our method outperforms others in restoring the best visual effects. Unlike alternative approaches that fail to learn uncertainty, leading to pattern degradation such as NAFNet (Chen et al., 2022) and MHNet (Gao and Dang, 2023), our technique successfully reconstructs images with superior visual quality.

Figure 5. Image restoration results on synthetic and real sand dust images from our proposed Sand11K dataset.

Figure 6. Demonstrating visual effects for scenes with high uncertainty. Methods incapable of learning uncertainty exhibit pattern degradation, whereas our approach showcases a robust capability for uncertainty estimation, resulting in the best visual restoration among the compared techniques.
In addition, to demonstrate the ability of our proposed method in other similar image restoration tasks, such as image dehazing and image raindrop Removal, we also conducted corresponding experiments using publicly available datasets (Liu et al., 2021; Qian et al., 2018). As shown in Table 3 and Figures 7, 8, our method also performs extremely well in image dehazing and raindrop Removal tasks.

Figure 7. Image restoration results on synthetic and real sand dust images from the public Haze4K (Liu et al., 2021) dataset.

Figure 8. Image restoration results on synthetic and real sand dust images from the public Raindrop-B (Qian et al., 2018) dataset.
4.3 Ablation study
Table 4 presents ablation experimental metrics for UPSFE, Deep Priors, CFEM, and EFS. For the section on ablation experiments, as shown in Table 5, we illustrate the importance of individual components of our method. Tables 4, 5 demonstrate that our proposed component enhances the entire model's performance, underscoring our proposed method's rationality.
5 Conclusion
In this study, we introduce for the first time the concept of uncertainty modeling in the realm of sand dust image recovery. Building upon this, we present the HIUNet framework, which integrates our innovative Efficient Frequency Selector, a Deep Prior Feature Extractor, and a Cross-scale Feature Enhancement Module. Extensive evaluations confirm our method's superiority over state-of-the-art image restoration techniques, with significant advantages in sand-affected scenes. For the future work, we will further advance research on uncertainty estimation modeling and its application in the field of low-level vision.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
BW: Writing – original draft, Writing – review & editing. HL: Writing – review & editing. CQ: Writing – review & editing. HS: Validation, Writing – review & editing. YC: Visualization, Writing – review & editing. YW: Supervision, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This research was supported by Guangxi Science and Technology Base and Talent Special Project: Research and Application of Key Technologies for Precise Navigation (Gui Ke AD25069103) and National Key R&D Program of China (2023YFB2504400).
Acknowledgments
Thanks for all the help of the teachers and students of the related universities.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Al-Ameen, Z. (2016). Visibility enhancement for images captured in dusty weather via tuned tri-threshold fuzzy intensification operators. Int. J. Intell. Syst. Technol. Appl. 8, 10–17. doi: 10.5815/ijisa.2016.08.02
Chang, J., Lan, Z., Cheng, C., and Wei, Y. (2020). “Data uncertainty learning in face recognition,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (Seattle, WA: IEEE), 5710–5719. doi: 10.1109/CVPR42600.2020.00575
Charbonnier, P., Blanc-Feraud, L., Aubert, G., and Barlaud, M. (1994). Two deterministic half-quadratic regularization algorithms for computed imaging. Proc. 1st Int. Conf. Image Process. 2, 168–172. doi: 10.1109/ICIP.1994.413553
Chen, L., Chu, X., Zhang, X., and Sun, J. (2022). “Simple baselines for image restoration,” in European Conference on Computer Vision (Springer: New York), 17–33. doi: 10.1007/978-3-031-20071-7_2
Everingham, M., Eslami, S. M. A., Van Gool, L., Williams, C. K. I., Winn, J., and Zisserman, A. (2015). The pascal visual object classes challenge: a retrospective. Int. J. Comput. Vis. 111, 98–136. doi: 10.1007/s11263-014-0733-5
Gao, H., and Dang, D. (2023). Mixed hierarchy network for image restoration. arXiv preprint arXiv:2302.09554. doi: 10.48550/arXiv.2302.09554
Gao, H., Yang, J., Zhang, Y., Wang, N., Yang, J., and Dang, D. (2024). “A novel single-stage network for accurate image restoration,” in Visual Computer (Springer). doi: 10.1007/s00371-024-03599-6
Guo, H., Wang, H., and Ji, Q. (2022). “Uncertainty-guided probabilistic transformer for complex action recognition,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (New Orleans, LA: IEEE), 20052–20061. doi: 10.1109/CVPR52688.2022.01942
He, K., Chen, X., Xie, S., Li, Y., Dollár, P., and Girshick, R. (2022). “Masked autoencoders are scalable vision learners,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (New Orleans, LA: IEEE), 16000–16009. doi: 10.1109/CVPR52688.2022.01553
He, K., Sun, J., and Tang, X. (2010). Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 33, 2341–2353. doi: 10.1109/TPAMI.2010.168
Hu, J., Shen, L., and Sun, G. (2018). “Squeeze-and-excitation networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Salt Lake City, UT: IEEE), 7132–7141. doi: 10.1109/CVPR.2018.00745
Jiang, K., Wang, Z., Yi, P., Chen, C., Huang, B., Luo, Y., et al. (2020). “Multi-scale progressive fusion network for single image deraining,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (Seattle, WA: IEEE), 8346–8355. doi: 10.1109/CVPR42600.2020.00837
Korb, K. B., and Nicholson, A. E. (2010). Bayesian Artificial Intelligence. CRC Press. doi: 10.1201/b10391
Lee, H., Kim, H.-E., and Nam, H. (2019). “SRM: a style-based recalibration module for convolutional neural networks,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (Seoul: IEEE), 1854–1862. doi: 10.1109/ICCV.2019.00194
Li, B., Liu, X., Hu, P., Wu, Z., Lv, J., and Peng, X. (2022). “All-in-one image restoration for unknown corruption,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (New Orleans, LA: IEEE), 17452–17462. doi: 10.1109/CVPR52688.2022.01693
Li, B., Peng, X., Wang, Z., Xu, J., and Feng, D. (2017a). An all-in-one network for dehazing and beyond. arXiv preprint arXiv:1707.06543. doi: 10.48550/arXiv.1707.06543
Li, Y., You, S., Brown, M. S., and Tan, R. T. (2017b). Haze visibility enhancement: a survey and quantitative benchmarking. Comput. Vis. Image Underst. 165, 1–16. doi: 10.1016/j.cviu.2017.09.003
Liu, Y., Zhu, L., Pei, S., Fu, H., Qin, J., Zhang, Q., et al. (2021). “From synthetic to real: image dehazing collaborating with unlabeled real data,” in Proceedings of the 29th ACM International Conference on Multimedia (Chengdu: ACM), 50–58. doi: 10.1145/3474085.3475331
Mittal, A., Moorthy, A. K., and Bovik, A. C. (2012a). No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 21, 4695–4708. doi: 10.1109/TIP.2012.2214050
Mittal, A., Soundararajan, R., and Bovik, A. C. (2012b). Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 20, 209–212. doi: 10.1109/LSP.2012.2227726
Narasimhan, S. G., and Nayar, S. K. (2002). Vision and the atmosphere. Int. J. Comput. Vis. 48, 233–254. doi: 10.1023/A:1016328200723
Neal, R. M. (2012). Bayesian Learning for Neural Networks, Volume 118. Springer Science and Business Media.
Peng, Y.-T., Cao, K., and Cosman, P. C. (2018). Generalization of the dark channel prior for single image restoration. IEEE Trans. Image Process. 27, 2856–2868. doi: 10.1109/TIP.2018.2813092
Qian, R., Tan, R. T., Yang, W., Su, J., and Liu, J. (2018). “Attentive generative adversarial network for raindrop removal from a single image,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Salt Lake City, UT: IEEE), 2482–2491. doi: 10.1109/CVPR.2018.00263
Qin, X., Wang, Z., Bai, Y., Xie, X., and Jia, H. (2020). Ffa-net: feature fusion attention network for single image dehazing. Proc. AAAI Conf. Artif. Intell. 34, 11908–11915. doi: 10.1609/aaai.v34i07.6865
Shi, J., Wei, B., Zhou, G., and Zhang, L. (2023). “Sandformer: CNN and transformer under gated fusion for sand dust image restoration,” in ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (Rhodes: IEEE), 1–5. doi: 10.1109/ICASSP49357.2023.10095242
Shi, Z., Feng, Y., Zhao, M., Zhang, E., and He, L. (2019). Let you see in sand dust weather: a method based on halo-reduced dark channel prior dehazing for sand-dust image enhancement. IEEE Access 7, 116722–116733. doi: 10.1109/ACCESS.2019.2936444
Si, Y., Yang, F., Guo, Y., Zhang, W., and Yang, Y. (2022). A comprehensive benchmark analysis for sand dust image reconstruction. J. Vis. Commun. Image Represent. 89:103638. doi: 10.1016/j.jvcir.2022.103638
Vaswani, A. (2017). “Attention is all you need,” in Advances in Neural Information Processing Systems (Long Beach, CA: Curran Associates, Inc.).
Wang, H., and Yeung, D.-Y. (2020). A survey on bayesian deep learning. ACM Comput. Surv. 53, 1–37. doi: 10.1145/3409383
Wang, Z., Li, Y., Guo, Y., Fang, L., and Wang, S. (2021). “Data-uncertainty guided multi-phase learning for semi-supervised object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (Nashville, TN: IEEE), 4568–4577. doi: 10.1109/CVPR46437.2021.00454
Woo, S., Park, J., Lee, J.-Y., and Kweon, I. S. (2018). “Cbam: convolutional block attention module,” in Proceedings of the European Conference on Computer Vision (ECCV) (Munich: Springer), 3–19. doi: 10.1007/978-3-030-01234-2_1
Wu, H., Qu, Y., Lin, S., Zhou, J., Qiao, R., Zhang, Z., et al. (2021). “Contrastive learning for compact single image dehazing,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (Nashville, TN: IEEE), 10551–10560. doi: 10.1109/CVPR46437.2021.01041
Yang, F., Zhai, Q., Li, X., Huang, R., Luo, A., Cheng, H., et al. (2021). “Uncertainty-guided transformer reasoning for camouflaged object detection.” in Proceedings of the IEEE/CVF International Conference on Computer Vision (Montreal, QC: IEEE), 4146–4155. doi: 10.1109/ICCV48922.2021.00411
Zamir, S. W., Arora, A., Khan, S., Hayat, M., Khan, F. S., Yang, M.-H., et al. (2021). “Multi-stage progressive image restoration,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (Nashville, TN: IEEE), 14821–14831. doi: 10.1109/CVPR46437.2021.01458
Zamir, S. W., Arora, A., Khan, S., Hayat, M., Khan, F. S., and Yang, M.-H. (2022). “Restormer: Efficient transformer for high-resolution image restoration,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (New Orleans, LA: IEEE), 5728–5739. doi: 10.1109/CVPR52688.2022.00564
Keywords: uncertainty-aware, image restoration, sand removal, hierarchical interaction, feature selection
Citation: Wei B, Liu H, Qian C, Shen H, Chen Y and Wang Y (2025) Toward accurate single image sand dust removal by utilizing uncertainty-aware neural network. Front. Neurorobot. 19:1575995. doi: 10.3389/fnbot.2025.1575995
Received: 13 February 2025; Accepted: 25 August 2025;
Published: 10 September 2025.
Edited by:
Jorge Dias, Khalifa University, United Arab EmiratesReviewed by:
Vidya Sudevan, Khalifa University, United Arab EmiratesZohair Al-Ameen, University of Mosul, Iraq
Copyright © 2025 Wei, Liu, Qian, Shen, Chen and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Hui Liu, bG93ZWxpdUB3aHUuZWR1LmNu