Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Physical model and image translation fused network for single-image dehazing

Published: 01 October 2023 Publication History

Highlights

An end-to-end physical model and image translate fused network is proposed for single image dehazing. The proposed network contains a common feature extraction encoder, a transmission map decoder and an image translation decoder.
An elaborate conditional attention feature fusion block is proposed to aggregate the guidance information of the physical model and the image translation.
The network is supervised with multi-task multi-scale deep supervision mechanism and verified on low-level image dehazing task and high-level semantic segmentation task.

Abstract

The visibility and contrast of images captured in adverse weather such as haze or fog degrade dramatically, which further hinders the accomplishment of high-level computer vision tasks such as object detection and semantic segmentation in these conditions. Many methods have been proposed to solve image dehazing problem by using image translation networks or physical model embedding in CNNs. However, the physical model cannot effectively describe the hazy generation process in complex scenes and estimating the model parameters with only a hazy image is an ill-posed problem. Image translation-based methods may lead to artefacts or colour shifts in the recovered results without the guidance or constraints of physical model information. In this paper, an end-to-end physical model and image translation fused network is proposed to generate realistic haze-free images. Since the transmission map can express the haze distribution in the scene, the proposed method adopts an encoder with a multiscale residual block to extract hazy image features, and two separate decoders to recover a clear image and to estimate the transmission map. The multiscale features of the transmission map and image translation are fused to guide the decode processes with a conditional attention feature fusion block, which is composed of sequential channelwise and spatialwise attention. Moreover, a multitask and multiscale deep supervision mechanism is adopted to enhance the feature fusion and recover more image details. The algorithm can efficiently fuse the physical model information and the hazy image translation to address the problem existent in the methods only based on physical model embedding or direct image translation. Experimental results on the visual quality enhancement of hazy images and semantic segmentation tasks in hazy scenes demonstrate that our model can efficiently recover haze-free images, while performing on par with state-of-the-art methods.

References

[1]
E.J. McCartney, “Optics of the atmosphere: scattering by molecules and particles,” New York, John Wiley and Sons, Inc., 1976. (1) 421.
[2]
S. Yin, Y. Wang, Y.H. Yang, A novel image-dehazing network with a parallel attention block, Pattern Recognit. 102 (2020).
[3]
F. Yuan, Y. Zhou, X. Xia, et al., A confidence prior for image dehazing, Pattern Recognit. 119 (2021).
[4]
K. He, J. Sun, X. Tang, Single image haze removal using dark channel prior, IEEE Trans. Pattern Anal. Mach. Intell. 33 (12) (2011) 2341–2353.
[5]
R. Fattal, Dehazing using color-lines, ACM Trans. Graph. TOG 34 (1) (2014) 1–14.
[6]
W. Ren, J. Pan, H. Zhang, et al., Single image dehazing via multi-scale convolutional neural networks with holistic edges, Int. J. Comput. Vis. 128 (1) (2020) 240–259.
[7]
H. Zhang, V.M. Pattel, Densely connected pyramid dehazing network, in: Proceedings of the Computer Society Conference on Computer Vision and Pattern Recognition, 2018, pp. 3194–3203.
[8]
J. Shen, Z. Li, L. Yu, et al., Implicit Euler ODE networks for single-image dehazing, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020, pp. 218–219.
[9]
H. Dong, J. Pan, L. Xiang, et al., Multi-scale boosted dehazing network with dense feature fusion, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 2157–2167.
[10]
Z. Chen, Y. Wang, Y. Yang, et al., PSD: principled synthetic-to-real dehazing guided by physical priors, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 7180–7189.
[11]
Y. Shao, L. Li, W. Ren, et al., Domain adaptation for image dehazing, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 2808–2817.
[12]
M. Hong, Y. Xie, C. Li, et al., Distilling image dehazing with heterogeneous task imitation, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 3462–3471.
[13]
S.K. Nayar, S.G. Narasimhan, Contrast restoration of weather degraded images, IEEE Trans. Pattern Anal. Mach. Intell. 25 (6) (2003) 713–724.
[14]
Y. Liu. J. Pan, J. Ren, Z. Su, Learning deep priors for image dehazing, in: Proceedings of the IEEE International Conference on Computer Vision, 2019, pp. 2492–2500.
[15]
Z. Deng, L. Zhu, X. Hu, C. Fu, Deep multi-model fusion for single-image dehazing, in: Proceedings of the IEEE International Conference on Computer Vision, 2019, pp. 2453–2462.
[16]
J. Dong, J. Pan, Physics-based feature dehazing networks, in: Proceedings of the European Conference on Computer Vision, 2020, pp. 188–204.
[17]
Q. Deng, Z. Huang, C. Tsai, et al., Hardgan: a haze-aware representation distillation gan for single image dehazing, in: Proceedings of the European Conference on Computer Vision, 2020, pp. 722–738.
[18]
S.M. Pizer, E.P. Amburn, J.D. Austin, et al., Adaptive histogram equalization and its variations, Comput. Vis. Graph. Image Process. 39 (3) (1987) 355–368.
[19]
R. Li, J. Pan, Z. Li, J. Tang, Single image dehazing via conditional generative adversarial network, in: Proceedings of the Computer Society Conference on Computer Vision and Pattern Recognition, 2018, pp. 8202–8211.
[20]
Y. Qu, Y. Chen, J. Huang, Y. Xie, Enhanced pix2pix dehazing network, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 8160–8168.
[21]
D. Chen, M. He, Q. Fan, et al., Gated context aggregation network for image dehazing and deraining, in: Proceedings of the IEEE winter conference on applications of computer vision (WACV), 2019, pp. 1375–1383. IEEE.
[22]
X. Liu, Y. Ma, Z. Shi, J. Chen, GridDehazeNet: attention-based multi-scale network for image dehazing, in: Proceedings of the IEEE International Conference on Computer Vision, 2019, pp. 7314–7323.
[23]
X. Qin, Z. Wang, Y. Bai, et al., FFA-Net: feature fusion attention network for single image dehazing, in: Proceedings of the AAAI Conference on Artificial Intelligence, 2020, pp. 11908–11915.
[24]
S. Yin, X. Yang, Y. Wang, et al., Visual attention dehazing network with multi-level features refinement and fusion, Pattern Recognit. 118 (2021).
[25]
N. Ibtehaz, M. Rahman, MultiResUNet: rethinking the U-Net architecture for multimodal biomedical image segmentation [J], Neural Netw. 121 (2020) 74–87.
[26]
L. Li, Y. Dong, W. Ren, et al., Semi-supervised image dehazing, IEEE Trans. Image Process. 29 (2019) 2766–2779.
[27]
T. Wang, M. Liu, J. Zhu, A. Tao, et al., Two-step domain adaptation for underwater image enhancement, Pattern Recognit. 122 (2022).
[28]
J. Hu, L. Shen, G. Sun, Squeeze-and-excitation networks, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 7132–7141.
[29]
B. Sun, X. Ye, B. Li, et al., Learning scene structure guidance via cross-task knowledge transfer for single depth super-resolution, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 7792–7801.
[30]
Z. Shen, Z. Liu, J. Li, et al., Dsod: learning deeply supervised object detectors from scratch, in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 1919–1927.
[31]
C.Y. Lee, S. Xie, P. Gallagher, et al., Deeply-supervised nets, in: Proceedings of the Artificial Intelligence and Statistics, 2015, pp. 562–570.
[32]
H. Zhao, O. Gallo, I. Frosio, et al., Loss functions for image restoration with neural networks, IEEE Trans. Image Process. 3 (1) (2016) 47–57.
[33]
D. Ren, W. Shang, P. Zhu, et al., Single image deraining using bilateral recurrent network, IEEE Trans. Image Process. (29) (2020) 6852–6863.
[34]
B. Li, W. Ren, D. Fu, et al., Benchmarking single-image dehazing and beyond, IEEE Trans. Image Process. 28 (1) (2019) 492–505.
[35]
C. Sakaridis, D. Dai, L. Van Gool, Semantic foggy scene understanding with synthetic data, Int. J. Comput. Vis. 126 (9) (2018) 973–992.
[36]
M. Cordts, M. Omran, S. Ramos, et al., The cityscapes dataset for semantic urban scene understanding, in: Proceedings of the IEEE conference on computer vision and Pattern Recognition, 2016, pp. 3213–3223.
[37]
Y. Zhang, L. Ding, G. Sharma, Hazerd: an outdoor scene dataset and benchmark for single image dehazing, in: Proceedings of the IEEE International Conference on Image Processing, 2017, pp. 3205–3209.
[38]
S. Zhao, L. Zhang, S. Huang, et al., Dehazing evaluation: real-world benchmark datasets, criteria, and baselines, IEEE Trans. Image Process. (29) (2020) 6947–6962.
[39]
D. Berman, S. Avidan, et al., Non-local image dehazing, in: Proceedings of the Computer Society Conference on Computer Vision and Pattern Recognition, 2016, pp. 1674–1682.
[40]
H. Zhao, J. Shi, X. Qi, et al., Pyramid scene parsing network, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 2881–2890.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Pattern Recognition
Pattern Recognition  Volume 142, Issue C
Oct 2023
767 pages

Publisher

Elsevier Science Inc.

United States

Publication History

Published: 01 October 2023

Author Tags

  1. Single-image dehazing
  2. Image translation
  3. Physical model
  4. Feature fusion

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 12 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2025)Dual-domain multi-scale feature extraction for image dehazingMultimedia Systems10.1007/s00530-024-01630-331:1Online publication date: 1-Feb-2025
  • (2025)Sat-DehazeGAN: an efficient dehazing model in water-sky background for river-sea transportMultimedia Systems10.1007/s00530-024-01599-z31:1Online publication date: 1-Feb-2025
  • (2024)Knowledge-guided multi-perception attention network for image dehazingThe Visual Computer: International Journal of Computer Graphics10.1007/s00371-023-03177-240:9(6479-6492)Online publication date: 1-Sep-2024
  • (2023)LCDA-Net: Efficient Image Dehazing with Contrast-Regularized and Dilated AttentionNeural Processing Letters10.1007/s11063-023-11384-055:8(11467-11488)Online publication date: 1-Dec-2023

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media