Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

EUICN: An Efficient Underwater Image Compression Network

Published: 26 February 2024 Publication History

Abstract

Thriving ocean applications bring explosive growth of underwater images (UWIs), which urgently demand to be compressed efficiently for transmission in the narrow underwater acoustic channel. However, existing image compression networks achieve suboptimal performance on UWIs. More efficient UWI compression can be achieved by utilizing characteristics of UWIs: (1) Within an UWI, the details distribution is associated with the underwater imaging transmission map (T-map); (2) Different UWIs have higher correlation than terrestrial images because they often present gauzy-covered indistinct appearance and share some universal ocean objects that widely appear in different underwater scenes. This paper fully exploits the two characteristics in terms of quantization and entropy coding, two key components of image compression network. Specifically, we propose an efficient underwater image compression network (EUICN) including underwater T-map-based quantization (UTMQ) and mixture entropy coding (MEC). In which, UTMQ extracts the imaging features from T-map, which are integrated with latent features by a novel dual-spatial attention module (DSAM) to generate a feature reserved mask, adaptively reserving reasonable numbers of features for different regions. For more efficient entropy coding, MEC is designed, which includes three correlation information extraction modules (<inline-formula> <tex-math notation="LaTeX">$i.e.$ </tex-math></inline-formula>, hyperprior, local and novel universal information) and a probability prediction module. Especially, the universal information extraction module utilizes a comprehensive underwater feature dictionary, which covers various universal ocean objects, to match with latent features to select the universal correlation features. After that, the probability prediction module is designed to consolidate the hyperprior, local, and universal information to predict more accurate probability of latent features. Extensive experiments show that our EUICN achieves better performance than SOTA learned and conventional codecs in terms of PSNR and MS-SSIM.

References

[1]
G. L. Foresti, “Visual inspection of sea bottom structures by an autonomous underwater vehicle,” IEEE Trans. Syst., Man, Cybern., B, vol. 31, no. 5, pp. 691–705, Oct. 2001.
[2]
J. Åhlén and D. Sundgren, “Bottom reflectance influence on a color correction algorithm for underwater images,” in Proc. Image Anal. Conf., 2003, pp. 922–926.
[3]
Y. Kahanov and J. G. Royal, “Analysis of hull remains of the Dor D Vessel, Tantura Lagoon, Israel,” Int. J. Nautical Archaeol., vol. 30, no. 2, pp. 257–265, Oct. 2001.
[4]
C.-Y. Chen, H.-A. Hou, and C.-M. Shu, “Hybrid-band relay based underwater video broadcasting systems with applications to entertainment and exploration,” in Proc. IEEE Int. Symp. Broadband Multimedia Syst. Broadcast. (BMSB), Mar. 2010, pp. 1–5.
[5]
X. Zhu, J. Song, L. Gao, F. Zheng, and H. T. Shen, “Unified multivariate Gaussian mixture for efficient neural image compression,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2022, pp. 17591–17600.
[6]
R. Zou, C. Song, and Z. Zhang, “The devil is in the details: Window-based attention for image compression,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2022, pp. 17471–17480.
[7]
Y. Xie, K. L. Cheng, and Q. Chen, “Enhanced invertible encoding for learned image compression,” in Proc. 29th ACM Int. Conf. Multimedia, Oct. 2021, pp. 162–170.
[8]
M. Rabbani, “JPEG2000: Image compression fundamentals, standards and practice,” J. Electron. Imag., vol. 11, no. 2, p. 286, 2002.
[9]
G. J. Sullivan, J.-R. Ohm, W.-J. Han, and T. Wiegand, “Overview of the high efficiency video coding (HEVC) standard,” IEEE Trans. Circuits Syst. Video Technol., vol. 22, no. 12, pp. 1649–1668, Dec. 2012.
[10]
B. Brosset al., “Overview of the versatile video coding (VVC) standard and its applications,” IEEE Trans. Circuits Syst. Video Technol., vol. 31, no. 10, pp. 3736–3764, Oct. 2021.
[11]
J. Y. Chiang and Y. Chen, “Underwater image enhancement by wavelength compensation and dehazing,” IEEE Trans. Image Process., vol. 21, no. 4, pp. 1756–1769, Apr. 2012.
[12]
D. Akkaynak, T. Treibitz, T. Shlesinger, Y. Loya, R. Tamir, and D. Iluz, “What is the space of attenuation coefficients in underwater computer vision?,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 568–577.
[13]
Q. Qi, K. Li, H. Zheng, X. Gao, G. Hou, and K. Sun, “SGUIE-Net: Semantic attention guided underwater image enhancement with multi-scale perception,” IEEE Trans. Image Process., vol. 31, pp. 6816–6830, 2022.
[14]
J. Ballé, D. Minnen, S. Singh, S. Hwang, and N. Johnston, “Variational image compression with a scale hyperprior,” in Proc. 6th Int. Conf. Learn. Rep. (ICLR), Apr. 2018, pp. 1–47.
[15]
D. Minnen, J. Ballé, and G. D. Toderici, “Joint autoregressive and hierarchical priors for learned image compression,” in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2018, pp. 10771–10780.
[16]
T. Chen, H. Liu, Z. Ma, Q. Shen, X. Cao, and Y. Wang, “End-to-end learnt image compression via non-local attention optimization and improved context modeling,” IEEE Trans. Image Process., vol. 30, pp. 3179–3191, 2021.
[17]
Z. Cheng, H. Sun, M. Takeuchi, and J. Katto, “Learned image compression with discretized Gaussian mixture likelihoods and attention modules,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2020, pp. 7936–7945.
[18]
F. Mentzer, E. Agustsson, M. Tschannen, R. Timofte, and L. V. Gool, “Conditional probability models for deep image compression,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018, pp. 4394–4402.
[19]
M. Li, W. Zuo, S. Gu, J. You, and D. Zhang, “Learning content-weighted deep image compression,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 43, no. 10, pp. 3446–3461, Oct. 2021.
[20]
J. Cai and L. Zhang, “Deep image compression with iterative non-uniform quantization,” in Proc. 25th IEEE Int. Conf. Image Process. (ICIP), Oct. 2018, pp. 451–455.
[21]
J. Lee, S. Cho, and S.-K. Beack, “Context-adaptive entropy model for end-to-end optimized image compression,” in Proc. 7th Int. Conf. Learn. Rep. (ICLR), May 2019, pp. 1–20.
[22]
Z. Guo, Y. Wu, R. Feng, Z. Zhang, and Z. Chen, “3-D context entropy model for improved practical image compression,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), Jun. 2020, pp. 520–523.
[23]
D. Minnen, G. Toderici, S. Singh, S. J. Hwang, and M. Covell, “Image-dependent local entropy models for learned image compression,” in Proc. 25th IEEE Int. Conf. Image Process. (ICIP), Oct. 2018, pp. 430–434.
[24]
Y. Qianet al., “Learning accurate entropy model with global reference for image compression,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2021, pp. 1–17.
[25]
Z. Guo, Z. Zhang, R. Feng, and Z. Chen, “Causal contextual prediction for learned image compression,” IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 4, pp. 2329–2341, Apr. 2022.
[26]
D. He, Z. Yang, W. Peng, R. Ma, H. Qin, and Y. Wang, “ELIC: Efficient learned image compression with unevenly grouped space-channel contextual adaptive coding,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2022, pp. 5708–5717.
[27]
Y. Qian, M. Lin, X. Sun, Z. Tan, and R. Jin, “Entroformer: A transformer-based entropy model for learned image compression,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2022, pp. 1–14.
[28]
Q.-Z. Li and W.-J. Wang, “Low-bit-rate coding of underwater color image using improved wavelet difference reduction,” J. Vis. Commun. Image Represent., vol. 21, no. 7, pp. 762–769, Oct. 2010.
[29]
Y. Zhang, Q.-Z. Li, and S. Negahdaripour, “Seafloor image compression using hybrid wavelets and directional filter banks,” in Proc. OCEANS, May 2015, pp. 1–9.
[30]
Y. He, X. Bu, M. Jiang, and M. Fan, “Low bit rate underwater video image compression and coding method based on wavelet decomposition,” China Commun., vol. 17, no. 9, pp. 210–219, Sep. 2020.
[31]
Y. Zhang, S. Negahdaripour, and Q. Li, “Low bit-rate compression of underwater imagery based on adaptive hybrid wavelets and directional filter banks,” Signal Process., Image Commun., vol. 47, pp. 96–114, Sep. 2016.
[32]
D. F. Hoag, V. K. Ingle, and R. J. Gaudette, “Low-bit-rate coding of underwater video using wavelet-based compression algorithms,” IEEE J. Ocean. Eng., vol. 22, no. 2, pp. 393–400, Apr. 1997.
[33]
J. Ahn, S. Yasukawa, T. Sonoda, Y. Nishida, K. Ishii, and T. Ura, “Image enhancement and compression of deep-sea floor image for acoustic transmission,” in Proc. OCEANS, Apr. 2016, pp. 1–6.
[34]
R. Monika, D. Samiappan, and R. Kumar, “Underwater image compression using energy based adaptive block compressive sensing for IoUT applications,” Vis. Comput., vol. 37, no. 6, pp. 1499–1515, Jun. 2021.
[35]
J. Wang, B. He, S. Zhang, R. Nian, Y. Shen, and T. Yan, “An efficient digital image compression scheme based on ICA for underwater color images,” in Proc. OCEANS, Apr. 2014, pp. 1–4.
[36]
N. Krishnaraj, M. Elhoseny, M. Thenmozhi, M. M. Selim, and K. Shankar, “Deep learning model for real-time image compression in Internet of Underwater Things (IoUT),” J. Real-Time Image Process., vol. 17, no. 6, pp. 2097–2111, Dec. 2020.
[37]
M. Zhuang, Y. Luo, X. Ding, Y. Huang, and Y. Liao, “A robustness and low bit-rate image compression network for underwater acoustic communication,” in Proc. Int. Conf. Neural Inf. Process., 2019, pp. 106–116.
[38]
M. Li, L. Shen, Y. Lin, K. Wang, and J. Chen, “Extreme underwater image compression using physical priors,” IEEE Trans. Circuits Syst. Video Technol., vol. 33, no. 4, pp. 1937–1951, Apr. 2023.
[39]
Y. Mei, L. Li, Z. Li, and F. Li, “Learning-based scalable image compression with latent-feature reuse and prediction,” IEEE Trans. Multimedia, vol. 24, pp. 4143–4157, 2022.
[40]
E. Agustssonet al., “Soft-to-hard vector quantization for end-to-end learning compressible representations,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), Dec. 2017, pp. 1141–1151.
[41]
L. Theis, W. Shi, A. Cunningham, and F. Huszár, “Lossy image compression with compressive autoencoders,” in Proc. 5th Int. Conf. Learn. Rep. (ICLR), Apr. 2017, pp. 1–19.
[42]
J. Ballé, V. Laparra, and E. P. Simoncelli, “End-to-end optimized image compression,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 20–46.
[43]
M. J. Islam, Y. Xia, and J. Sattar, “Fast underwater image enhancement for improved visual perception,” IEEE Robot. Autom. Lett., vol. 5, no. 2, pp. 3227–3234, Apr. 2020.
[44]
C. Liet al., “An underwater image enhancement benchmark dataset and beyond,” IEEE Trans. Image Process., vol. 29, pp. 4376–4389, 2020.
[45]
M. J. Islamet al., “Semantic segmentation of underwater imagery: Dataset and benchmark,” in Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst. (IROS), Oct. 2020, pp. 1769–1776.
[46]
O. Russakovskyet al., “ImageNet large scale visual recognition challenge,” Int. J. Comput. Vis., vol. 115, no. 3, pp. 211–252, Dec. 2015.
[47]
Y. Lin, L. Shen, Z. Wang, K. Wang, and X. Zhang, “Attenuation coefficient guided two-stage network for underwater image restoration,” IEEE Signal Process. Lett., vol. 28, pp. 199–203, 2021.
[48]
R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2018, pp. 586–595.
[49]
M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “GANs trained by a two time-scale update rule converge to a local Nash equilibrium,” in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2017, pp. 6629–6640.
[50]
Q. Jiang, Y. Gu, C. Li, R. Cong, and F. Shao, “Underwater image enhancement quality evaluation: Benchmark dataset and objective metric,” IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 9, pp. 5959–5974, Sep. 2022.
[51]
L. Hong, X. Wang, G. Zhang, and M. Zhao, “USOD10K: A new benchmark dataset for underwater salient object detection,” IEEE Trans. Image Process., early access, Apr. 14, 2023. 10.1109/TIP.2023.3266163.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image IEEE Transactions on Circuits and Systems for Video Technology
IEEE Transactions on Circuits and Systems for Video Technology  Volume 34, Issue 7
July 2024
1398 pages

Publisher

IEEE Press

Publication History

Published: 26 February 2024

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 0
    Total Downloads
  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 01 Jan 2025

Other Metrics

Citations

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media