Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Statistical Analysis of the Relationship between AETA Electromagnetic Anomalies and Local Earthquakes
Next Article in Special Issue
Unsupervised Exemplar-Domain Aware Image-to-Image Translation
Previous Article in Journal
Granger Causality on forward and Reversed Time Series
Previous Article in Special Issue
A Generative Adversarial Network for Infrared and Visible Image Fusion Based on Semantic Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Noisy SAR Image Fusion Method Based on NLM and GAN

1
Shandong Province Key Laboratory of Medical Physics and Image Processing Technology, School of Physics and Electronics, Shandong Normal University, Jinan 250014, China
2
Shandong Provincial Engineering and Technical Center of Light Manipulations & Shandong Provincial Key Laboratory of Optics and Photonic Device, School of Physics and Electronics, Shandong Normal University, Jinan 250014, China
3
Institute of Information Science, Beijing Jiaotong University, Beijing 100044, China
4
School of Environment Science and Spatial Informatics, China University of Mining and Technology, Xuzhou 221116, China
*
Authors to whom correspondence should be addressed.
Entropy 2021, 23(4), 410; https://doi.org/10.3390/e23040410
Submission received: 15 January 2021 / Revised: 26 March 2021 / Accepted: 27 March 2021 / Published: 30 March 2021
(This article belongs to the Special Issue Advances in Image Fusion)

Abstract

:
The unavoidable noise often present in synthetic aperture radar (SAR) images, such as speckle noise, negatively impacts the subsequent processing of SAR images. Further, it is not easy to find an appropriate application for SAR images, given that the human visual system is sensitive to color and SAR images are gray. As a result, a noisy SAR image fusion method based on nonlocal matching and generative adversarial networks is presented in this paper. A nonlocal matching method is applied to processing source images into similar block groups in the pre-processing step. Then, adversarial networks are employed to generate a final noise-free fused SAR image block, where the generator aims to generate a noise-free SAR image block with color information, and the discriminator tries to increase the spatial resolution of the generated image block. This step ensures that the fused image block contains high resolution and color information at the same time. Finally, a fused image can be obtained by aggregating all the image blocks. By extensive comparative experiments on the SEN1–2 datasets and source images, it can be found that the proposed method not only has better fusion results but is also robust to image noise, indicating the superiority of the proposed noisy SAR image fusion method over the state-of-the-art methods.

1. Introduction

As one of the active microwave imaging radars, synthetic aperture radar (SAR) can work at any time and in any weather conditions. The many advantages of SAR includes, among others, multi-polarization and variable angles, which allows SAR images to be widely used in geological surveys, military exercises, etc. [1,2]; however, due to its special coherent imaging mechanism, noise is inevitably generated in image acquisition, especially for speckle noise, resulting in serious inconvenience to the subsequent interpretation of the image processing; therefore, the effective suppression or removal of noise is one of the essential tasks required for SAR image pre-processing [3]. SAR can penetrate the earth’s surface as well as natural vegetation coverings, clearly and exhaustively map topography and geomorphology, and obtain high-resolution images of the earth’s surface; however, the color information of SAR images is relatively simple, and cannot adequately reflect the scene’s spectral information. On the contrary, multi-spectral sensors can obtain images with rich spectral information, such as color optical images [4]. Image fusion [5,6,7] is a powerful image processing tool for integrating complementary information from different sensors, by which a fused image with a more comprehensive and clearer description of the scene can be obtained. Although an increasing number of papers about image fusion are published every year—indicating the importance of image fusion—few papers are published regarding noisy SAR image fusion, despite the urgent need for an effective and practical SAR image fusion method.
Image fusion can be classified into pixel-level fusion, feature-level fusion, and decision-level fusion [8,9]. Pixel-level fusion fuses the pixels of source images directly, which is the basis of other level fusions; however, a significant amount of information has to be processed. Feature-level fusion extracts feature information of images such as edge, shape, texture, etc., then fuses them together; because it only extracts features for image fusion, detailed information is often missing. The most advanced decision-level fusion is based on feature-level fusion. After feature extraction, other image processing methods, including classification, recognition, and comprehensive evaluation, are employed to make a final decision. This kind of method is based on a cognitive model, which needs large databases and expert decisions for analysis.
Normally, pixel-level image fusion methods can be classified into five categories [6], including methods based on spatial domain, methods based on transform domain, methods based on sparse representation, methods based on deep learning, and hybrid methods. The methods based on spatial domain use some mathematical operations, such as “choose-max” and “average-weighting” to process the pixel value of source images directly and obtain the pixel value at the corresponding position of the fused image. The classic methods include image fusion based on maximization [10], image fusion based on intensity-hue-saturation (IHS) transform, principal component analysis (PCA) transform [11], and so on. To make full use of the image features, we employ image transforms to filter images into the feature domain. Image fusion can be realized in the transform domain. Similar to image transforms, sparse representation also transfers source images from the spatial domain to another image feature domain by the mathematical transformation; thus, this can be considered a special method based on image transform. In order to make full use of the characteristics of images in the spatial domain and the sparse features in the transform domain, many scholars mixed different kinds of image fusion methods together to obtain new image fusion methods, including image fusion based on the spatial domain and transform domain, image fusion based on the transform domain and sparse representation, and so on; however, these traditional methods usually involve fusion rule designing, which is difficult to adapt to the characteristics of different source images and cost much computing time [12].
In recent years, rapid progress has been made in deep learning, computer vision, and image processing [13], and a significant amount of research is currently being conducted in image fusion. For example, Liu [14] proposed a multi-focus image fusion method with a deep convolutional neural network; Zhong [15] proposed a remote sensing image fusion method with a convolutional neural network; Ma [12] is the first one who applied generative adversarial networks (GAN) into infrared and visible image fusion and achieved good fusion results; however, the construction of the existing FusionGAN-based method is simple, and loss fusion is imperfect, which can lead to incomplete information transfer in the fused image. Improved GAN-based methods have been proposed for image fusion, such as Zhang [16], who proposed a new generative adversarial network with adaptive and gradient joint constraints to fuse multi-focus images. Nevertheless, the methods mentioned above are effective for other kinds of source images rather than remote sensing images, and they are sensitive to the noise in an image. In this paper, we propose a noisy SAR image fusion method based on nonlocal matching (NLM) and GAN, which is more robust to noise and can serve more information from source images. The nonlocal theory is to find similar blocks in the whole image rather than in a local window and has been widely used in SAR image de-noising [17,18,19,20,21]. The nonlocal matching takes advantage of the existence of a pattern or similar features in the non-adjacent pixels and exploits self-similarities in the search neighborhood to estimate the true value of the noisy pixel. In this paper, NLM is employed on source images to acquire similar image block groups. The generator of GAN can generate a fused noise-free image block group, and the final fused image can be obtained by aggregating the blocks after the generator and discriminator are dynamically balanced. The main contributions of this paper can be summarized as follows:
(1)
Due to the existence of speckle noise, SAR image de-noising is a necessary pre-processing technology; however, in our approach, we develop SAR image de-noising and fusion simultaneously, which can avoid the complex pre-processing and save more time;
(2)
Nonlocal matching is employed as a pre-processing technology for GAN to obtain similar block groups, which makes full use of similar information in source images and provides more effective inputs for GAN;
(3)
For image fusion, “standard, well-integrated” reference images often do not exist; i.e., when a deep learning method is used to fuse the source images, there is no reference tag; therefore, GAN is employed to perform the image de-noising and fusion without reference images by limiting the loss functions.
The rest of this paper can be summarized as follows. In Section 2, the conceptualization of GAN and its family are described. Section 3 presents the proposed noisy SAR image fusion method. Section 4 gives more details about the comparative experimental results and analysis. Finally, the conclusion is made in Section 5.

2. GAN

Since GAN was proposed by Ian Goodfellow in 2014 [22], it has been widely used in image processing and other fields, such as GAN-based synthetic medical image augmentation [23], realistic image synthesis with stacked generative adversarial networks [24], and so on. In addition, GAN is also favored by researchers for image fusion. For example, Ma proposed Pan-GAN in remote sensing image fusion, which can obtain multi-spectral images of high-resolution by fusing panchromatic images and multispectral images of low-resolution [25], Li proposed coupled GAN with relativistic discriminators for infrared and visible image fusion, where the simple yet efficient relativistic discriminator is applied [26], and so on.
GAN, shown in Equation (1), is composed of two competing neural networks: a discriminator and a generator, where the G tries to generate data being close to the distribution P d a t a of real data, and the discriminator D tries to distinguish between real data and data generated by the generator. During the two network confrontations, the generator uses the discriminator as the loss function and updates its parameters to produce more realistic-looking data. On the other hand, the discriminator updates its generator parameters to better identify false data from real data; it relies on the standard multilayer perceptron architecture to realize the network.
min G max D V ( D , G ) = E x ~ P d a t a ( x ) log D ( x ) + E z ~ P z ( z ) log 1 D G ( z )
Subsequently, convolutional operations are introduced into GAN and the deep convolutional GAN (DCGAN) is generated. Sometimes, the loss of GAN fluctuates because the generator and discriminator undermine each other’s learning. As a result, the progressive growing of GAN (ProGAN) is a network that stabilizes GAN training by increasing the resolution of the generated images. Because of its transformation architecture, self-attention GAN (SAGAN) has also become much popular in recent years. Instead of focusing on creating more realistic images and improving the ability of GAN to perform fine control over the generated images, StyleGAN can be employed with any GAN to produce better results.

3. The Proposed Method

In this section, we describe the proposed noisy SAR image fusion method based on NLM and GAN. NLM is introduced first. Then, the network architecture of the proposed method is given to make it clearer and more readable.

3.1. NLM

NLM consists of image blocking and similarity grouping. To balance the effect and efficiency of image blocking, we abandon high-level blocking methods such as methods for image edges or regional features and adopt the traditional fixed-size sliding window blocking method. Some classical methods mainly use a certain distance between image blocks as a similar measurement. For two image blocks, the smaller their distance value is, the more similar they are. Common distance calculation methods include the method based on singular value, Euclidean distance, l p n o r m and so on. Euclidean distance is employed in this paper.

3.2. The Network of the Proposed Method

In order to retain the high spatial resolution of SAR image and color information of optical image more completely at the same time, we proposed a noisy SAR image fusion method based on NLM and GAN—shown in Figure 1. The similar image block groups obtained by NLM are fed into the generator at first, and the fused noise-free image I f can be generated by network training.
After that, the fused image and noisy SAR image I s are fed into the discriminator, which aims to determine whether the spatial resolution of fused and SAR images is consistent. More details about the network of the generator and discriminator are shown in Table 1.
At last, the loss function in Equation (2) of the proposed network contains two parts: the loss function of the generator L G in Equation (3), and the loss of the discriminator L D in Equation (5).
L o s s = a L G + b L D
where a and b are the weight factors that balance the contributions of the loss function of the generator and discriminator.
L G = L PSNR ( I f ) + l 2 ( I f , I o ) = 1 n 1 PSNR ( B f ) c o n s t + I f I o 2
where PSNR ( B f ) denotes peak signal-to-noise ratio (PSNR) of the fused image.
PSNR is often used to measure the noise level in an image:
PSNR ( B f ) = 10 · log 10 B max 2 MSE
The larger the PSNR, the better the image quality. c o n s t denotes a constant normalizing the value of PSNR and is set to 35, which means the maximum value of PSNR in an image, and the first term in Equation (3) can ensure that the fused image contains less noise; l 2 ( I f , I o ) denotes l 2 n o r m loss of the optical image I o and fused image.
L D = L SSIM ( I f , I s ) + l 2 ( I f , I s ) = 1 n 1 SSIM ( B f , B s ) + I f I s 2
where n denotes the total number of image blocks in source images. SSIM ( B f , B s ) denotes the structural similarity (SSIM) index of image blocks in the fused image and SAR image, which can be calculated as
SSIM ( B f , B s ) = 2 E [ B f ] · E [ B s ] + C 1 E [ B f 2 ] + E [ B s 2 ] + C 1 · 2 cov [ B f , B s ] + C 2 Var [ B f ] + Var [ B s ] + C 2
where B represents the image block, B f , B s represent the fused image block and the SAR image block, respectively. C 1 and C 2 denote constants that are not zero. SSIM is a number greater than 0 and less than 1, which measures the correlation loss, brightness loss and contrast loss between source images and the fused image. The closer SSIM is to 1, the more similar the structure is.

4. Experimental Results and Analysis

4.1. Datasets and Parameter Settings

The training datasets were selected from SEN1–2 datasets [4], which contain more than twenty hundred thousand SAR-optical image pairs with the size of 256 256 collected from across the globe and throughout all meteorological seasons. SAR images acquired by Sentinel-1 are polluted by speckle noise, whereas optical images acquired by Sentinel-2 are noise-free.
When NLM is employed to process the source images, the size of the image block is set to 32 32 , and the maximum image block in each similar group is 20. When training the proposed network, the generator and discriminator are optimized alternately, and we implemented our network in TensorFlow.

4.2. Compared Methods

To effectively evaluate the proposed noisy SAR image fusion method, in this section, we conducted the compared experiments by 8 representative image fusion methods, including image fusion based on guided filtering (GFF) [27], image fusion based on the sparse model (SR) [28], wavelet-based image fusion (DWT) [29], image fusion with deep convolutional neural network (CNN) [14], multi-scale weighted gradient-based fusion (MWGF) [30], image fusion based on multi-scale transform and sparse representation (MST-SR) [31], image fusion method in nonsubsampled Shearlet transform domain (NSST) [32], and a generative adversarial network for image fusion (GAN) [12]. Among all, MWGF and GFF belong to the methods based on spatial domain, while DWT and NSST are representative methods based on transform. For NSST, the employed fusion rule is “choose-max”. MST-SR is a hybrid method that combines diverse image fusion methods to implement better fused results, and NSCT is employed as multi-scale transform in this paper. CNN and GAN are popular methods based on deep learning. The codes of image fusion methods could be downloaded from links in their corresponding papers, and the parameters were set as recommended.

4.3. Valuable Metrics

To objectively evaluate different fusion methods, some objective metrics are employed to calculate the corresponding values of fused images, such as entropy (EN), average gradient (AVG), spatial frequency (SF), mutual information (MI), and QAB/F [33,34]. The larger these metric values are, the better the fused image is. The calculations of these metrics are as follows:
(1)
EN
When we want to measure how much information an image contains, EN [33] is a good choice, and it can reflect the average amount of information contained in the fused image. It can be calculated by Equation (7).
EN = i = 0 L 1 p i × log 2 p i
where L denotes the total number of pixels in an image, and p i is the probability distribution for pixels in each gray level.
(2)
AVG
The calculation of AVG is shown in Equation (8). We can evaluate the ability to vary tiny details and texture features in an image by the value of AVG.
AVG = 1 M N i = 1 M j = 1 N Δ I x 2 + Δ I y 2 2
where Δ I x = f ( x , y ) f ( x 1 , y ) and Δ I y = f ( x , y ) f ( x , y 1 ) .
(3)
SF
SF is used to detect the total activity of a fused image in the spatial domain, which represents the ability to contrast small details. It can be calculated by Equation (9).
SF ( i , j ) = ( RF ) 2 + ( CF ) 2
where RF = 1 M × N i = 1 M j = 2 N f ( x , y ) f ( x , y 1 ) 2 represents row frequency while CF = 1 M × N i = 2 M j = 1 N f ( x , y ) f ( x 1 , y ) 2 represents column frequency.
(4)
MI
MI represents the amount of information in a fused image from the source images, which also means the amount of information transferring from the source images to the fused image. More details about it are shown in Equation (10).
MI = MI AF + MI BF
where MI AF = f , a P FA ( f , a ) log P FA ( f , a ) P F ( f ) P A ( a ) and MI BF = f , b P FB ( f , b ) log P FB ( f , b ) P F ( f ) P B ( b ) . P FA ( f , a ) and P FB ( f , b ) denote the joint probability densities between the fused image F and the source images A,B, respectively, whereas P A ( a ) and P B ( a ) denote the probability densities of the source images.

4.4. Results and Analysis

4.4.1. Experiments on SEN1–2

We randomly selected thirty thousand pairs of source images to perform the experiments using our proposed network. Twenty thousand of them were used for the training set, whereas the other ten thousand were used for the validation set. Figure 2 shows some examples from the SEN1–2 datasets, which include images sourced from different seasons. The first column in Figure 2 is the SAR source images with much speckle noise. The second column in Figure 2 is the optical source images with rich color information. The third column is the fused images by the proposed method. From the snow in Figure 2k, we can conclude that the image was taken in winter; however, it is hard to infer the season only from its corresponding SAR image, which has high resolution but no spectral information. By comparing the source images to the optical images, the spatial resolution of fused images has significantly improved, meaning the proposed method can extract spatial information from noisy SAR images and color information from optical images simultaneously.
In order to better verify the proposed method in this paper, we performed comparative experiments on SEN1–2, and the source images are shown in Figure 3. To ensure the fairness of the experiments, image de-noising by SAR-BM3D [20] was done on noisy images for subsequent image fusion in the compared methods. The fused images of Groups 1~4 in Figure 3 are shown in Figure 4, Figure 5, Figure 6 and Figure 7. We can see that there are significant color distortions and black areas of the green lake in Figure 4a,e,g. The fused images in Figure 6f,i are easier to distinguish due to the appropriate coloring. By carefully comparing the details of the fused images in Group 4, it is obvious that the fused image in Figure 7i by the proposed noisy SAR image fusion method has less speckle noise. All in all, from these fused images, it can be found that the proposed noisy SAR image fusion method not only has better fusion results but also is robust to image noise.
Moreover, valuable metrics are employed to evaluate the fused images. To make these values clearer, we show the valuable metrics in Figure 8. By comparing these values in Figure 8, we can conclude that the values in green are larger than the others, showing that the images in Group 2 are better than the other groups; however, when comparing the fused images in the same group by different fusion methods, it can be found that the values of the fused images by the proposed method are better, in general, indicating that the proposed method has more power on SAR image fusion. To further illustrate the generalization of the proposed method, we continue by testing 10 groups of images in SEN1–2 datasets randomly, and the average objective indicators are shown in Table 2. From Table 2, we can see that the proposed method can also obtain higher objective indicators.

4.4.2. Experiments on Oslo City

To verify the superiority and practicability of the proposed method, we compared experiments on source images of Oslo city, which are noise-free—see Figure 9.
The fused images of Figure 9 by different image fusion methods are shown in Figure 10. There is some color distortion in Figure 10a,d,h, where the color of the forests is bright green, whereas the color in the optical image is dark green. Besides, due to the multi-scale transformation of the images, some detailed information of source images was missed when fusing them using DWT, MST-SR, and NSST—see Figure 10c,f,g. By comparing the fused images in Figure 10, the fused image by the proposed method in Figure 10i has a better subjective effect and is more visually suitable to the human eye.
To evaluate the fused images in Figure 10 more objectively, objective metrics and computational time were employed, and the results are shown in Table 3. Although the values of the fused image by GAN are near to ours in terms of EN, AVG, and SF, the value of MI in Figure 10h is less than two. From Table 3, we can conclude that the fused image by the proposed method has better values of objective metrics and costs less time than most of the other methods.

5. Conclusions

Making full use of similar structural features in an image, a robust noisy SAR image fusion method based on NLM and GAN is proposed in this paper. Using the adversarial game of the generator and discriminator, a final fused noise-free image with high spatial resolution and color information can be obtained, where the optimization of the proposed network is realized by the constraint of the constructed loss function in this paper. By comparing our experiments with the state-of-the-art image fusion methods on the SEN1–2 datasets and Oslo city, we demonstrated that the proposed noisy SAR image fusion is robust to image noise and has a better fusion effect, which makes the images more suitable for the human eye. The fusion images obtained by the proposed method have less residual noise and color distortion. Meanwhile, it can retain the edge and texture details of the source images more effectively compared with other methods. In conclusion, the proposed method has excellent image de-noising performance and a better fusion effect. It is an excellent image fusion method for noisy SAR images and optic images, which can be extended to the multi-sensor image fusion.
However, the available datasets are limited and the trained model relies on the trained datasets, which means that our results are difficult to generalize. In the future, we will explore or create more datasets and test the noisy image fusion model further to improve our method and obtain more ideal effects.

Author Contributions

Conceptualization, S.H. and Y.Z.; methodology, J.F.; software, X.M.; validation, J.F., X.M. and J.W.; formal analysis, J.F.; investigation, J.W.; resources, X.M.; data curation, J.F.; writing—original draft preparation, J.F.; writing—review and editing, J.F. and X.M.; visualization, X.M.; supervision, K.Q.; project administration, S.H. and Y.Z.; funding acquisition, J.F., S.H. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of China under grant Nos. 62002208 and 61572063: and Natural Science Foundation of Shandong Province under grant Nos. ZR2020MA082 and ZR2020MF119.

Data Availability Statement

Not applicable.

Acknowledgments

Source images in SEN1–2 dataset adopted in the experiments can be downloaded from the website of https://mediatum.ub.tum.de/1436631. We would love to thank Jiayi Ma, Zhiqin Zhu, Gonzalo Pajares, Yu Liu and so on for their shared codes of the compared experiments. We also thank the Editor and Reviewers for the efforts made in processing this submission, and we are particularly grateful to the reviewers for their constructive comments and suggestions, which helped us improve the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, Z.; Wang, H.; Xu, F.; Jin, Y.Q. Complex-valued convolutional neural network and its application in polarimetric SAR image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7177–7188. [Google Scholar] [CrossRef]
  2. Shen, H.; Lin, L.; Li, J.; Yuan, Q.; Zhao, L. A residual convolutional neural network for polarimetric SAR image super-resolution. Isprs J. Photogramm. Remote Sens. 2020, 161, 90–108. [Google Scholar] [CrossRef]
  3. Singh, P.; Shree, R. A new homomorphic and method noise thresholding based despeckling of SAR image using anisotropic diffusion. J. King Saud Univ. Comput. Inf. Sci. 2020, 3, 137–148. [Google Scholar] [CrossRef]
  4. Schmitt, M.; Hughes, L.H.; Körner, M.; Zhu, X.X. Colorizing sentinel-1 sar images using a variational autoencoder conditioned on sentinel-2 imagery. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 42, 1045–1051. [Google Scholar] [CrossRef] [Green Version]
  5. Liu, Y.; Chen, X.; Wang, Z.; Wang, Z.J.; Ward, R.K.; Wang, X. Deep learning for pixel-level image fusion: Recent advances and future prospects. Inf. Fusion 2018, 42, 158–173. [Google Scholar] [CrossRef]
  6. Li, S.; Kang, X.; Fang, L.; Hu, J.; Yin, H. Pixel-level image fusion: A survey of the state of the art. Inf. Fusion 2017, 33, 100–112. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Bai, X.; Wang, T. Boundary finding based multi-focus image fusion through multi-scale morphological focus-measure. Inf. Fusion 2017, 35, 81–101. [Google Scholar] [CrossRef]
  8. Yao, Y.F.; Jing, X.Y.; Wong, H.S. Face and palmprint feature level fusion for single sample biometrics recognition. Neurocomputing 2007, 70, 1582–1586. [Google Scholar] [CrossRef]
  9. Prabhakar, S.; Jain, A.K. Decision-level fusion in fingerprint verification. Pattern Recognit. 2002, 35, 861–874. [Google Scholar] [CrossRef]
  10. Mahyari, A.G.; Yazdi, M. Panchromatic and multispectral image fusion based on maximization of both spectral and spatial similarities. IEEE Trans. Geosci. Remote Sens. 2011, 49, 1976–1985. [Google Scholar] [CrossRef]
  11. He, C.; Liu, Q.; Li, H. Multimodal medical image fusion based on IHS and PCA. Procedia Eng. 2010, 7, 280–285. [Google Scholar] [CrossRef] [Green Version]
  12. Ma, J.; Yu, W.; Liang, P. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  13. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Liu, Y.; Chen, X.; Peng, H.; Wang, Z.F. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  15. Zhong, J.; Yang, B.; Huang, G.; Zhong, F.; Chen, Z. Remote sensing image fusion with convolutional neural network. Sens. Imaging 2016, 17, 10. [Google Scholar] [CrossRef]
  16. Zhang, H.; Le, Z.; Shao, Z.; Xu, H.; Ma, J. MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 2021, 66, 40–53. [Google Scholar] [CrossRef]
  17. Coupé, P.; Hellier, P.; Kervrann, C.; Barillot, C. Bayesian non local means-based speckle filtering. IEEE Int. Symp. Biomed. Imaging Nano Macro 2008, 1291–1294. [Google Scholar]
  18. Deledalle, C.; Denis, L.; Tupin, F. Iterative weighted maximum likelihood de-noising with probabilistic patch-based weights. IEEE Trans. Image Process. 2009, 18, 2661–2672. [Google Scholar] [CrossRef] [Green Version]
  19. Parrilli, S.; Poderico, M.; Angelino, C.; Scarpa, G.; Verdoliva, L. A nonlocal approach for SAR image de-noising. Geosci. Remote Sens. Symp. 2010, 45, 726–729. [Google Scholar]
  20. Parrilli, S.; Poderico, M.; Angelino, C.V.; Verdoliva, L. A nonlocal SAR image de-noising algorithm based on LLMMSE wavelet shrinkage. IEEE Trans. Geosci. Remote Sens. 2012, 50, 606–616. [Google Scholar] [CrossRef]
  21. Torres, L.; Frery, A.C. SAR image despeckling algorithms using stochastic distances and nonlocal means. arxiv 2013, arXiv:1308.4338. [Google Scholar]
  22. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 2672–2680. [Google Scholar]
  23. Frid-Adar, M.; Diamant, I.; Klang, E. GAN-based Synthetic Medical Image Augmentation for increased CNN Performance in Liver Lesion Classification. Neurocomputing 2018, 321, 321–331. [Google Scholar] [CrossRef] [Green Version]
  24. Han, Z.; Tao, X.; Hongsheng, L. StackGAN++: Realistic image synthesis with stacked generative adversarial networks. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 1947–1962. [Google Scholar]
  25. Ma, J.; Yu, W.; Chen, C. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  26. Li, Q.; Lu, L.; Li, Z. Coupled GAN with relativistic discriminators for infrared and visible images fusion. IEEE Sensors J. 2019. [Google Scholar] [CrossRef]
  27. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  28. Zhu, Z.; Yin, H.; Chai, Y.; Li, Y.; Qi, G. A novel multi-modality image fusion method based on image decomposition and sparse representation. Inf. Sci. 2018, 432, 516–529. [Google Scholar] [CrossRef]
  29. Pajares, G.; de la Cruz, J.M. A wavelet-based image fusion tutorial. Pattern Recognit. 2004, 37, 1855–1872. [Google Scholar] [CrossRef]
  30. Zhou, Z.; Li, S.; Wang, B. Multi-scale weighted gradient-based fusion for multi-focus images. Inf. Fusion 2014, 20, 60–72. [Google Scholar] [CrossRef]
  31. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  32. Yin, M.; Liu, W.; Zhao, X.; Yin, Y.; Guo, Y. A novel image fusion algorithm based on nonsubsampled shearlet transform. Opt. Int. J. Light Electron Opt. 2014, 125, 2274–2282. [Google Scholar] [CrossRef]
  33. Qu, X.; Yan, J.; Xiao, H.; Zhu, Z. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Autom. Sin. 2008, 34, 1508–1514. [Google Scholar] [CrossRef]
  34. Cui, Y. No-Reference Image Quality Assessment Based on Dual-Domain Feature Fusion. Entropy 2020, 22, 344. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. The network of the proposed method.
Figure 1. The network of the proposed method.
Entropy 23 00410 g001
Figure 2. Results of our experiments using the SEN1–2 datasets by the proposed method. The first column: noisy synthetic aperture radar (SAR) images; the second column: optical images; the third column: fused images. (ac) Group 1; (df) Group 2; (gi) Group 3; (jl) Group 4.
Figure 2. Results of our experiments using the SEN1–2 datasets by the proposed method. The first column: noisy synthetic aperture radar (SAR) images; the second column: optical images; the third column: fused images. (ac) Group 1; (df) Group 2; (gi) Group 3; (jl) Group 4.
Entropy 23 00410 g002
Figure 3. Compared source images in SEN1–2: the first column contains noisy SAR images, the second column contains de-noised SAR images, and the third column contains optical images. (a) Group 1; (b) Group 2; (c) Group 3; (d) Group 4.
Figure 3. Compared source images in SEN1–2: the first column contains noisy SAR images, the second column contains de-noised SAR images, and the third column contains optical images. (a) Group 1; (b) Group 2; (c) Group 3; (d) Group 4.
Entropy 23 00410 g003aEntropy 23 00410 g003b
Figure 4. The fused images of Group 1 in Figure 3: (a) guided filtering (GFF); (b) sparse model (SR); (c) wavelet-based image fusion (DWT); (d) convolutional neural network (CNN); (e) multi-scale weighted gradient-based fusion (MWGF); (f) multi-scale transform and sparse representation (MST-SR); (g) nonsubsampled Shearlet transform domain (NSST); (h) generative adversarial network (GAN); (i) the proposed method.
Figure 4. The fused images of Group 1 in Figure 3: (a) guided filtering (GFF); (b) sparse model (SR); (c) wavelet-based image fusion (DWT); (d) convolutional neural network (CNN); (e) multi-scale weighted gradient-based fusion (MWGF); (f) multi-scale transform and sparse representation (MST-SR); (g) nonsubsampled Shearlet transform domain (NSST); (h) generative adversarial network (GAN); (i) the proposed method.
Entropy 23 00410 g004
Figure 5. The fused images of Group 2 in Figure 3: (a) GFF; (b) SR; (c) DWT; (d) CNN; (e) MWGF; (f) MST-SR; (g) NSST; (h) GAN; (i) the proposed method.
Figure 5. The fused images of Group 2 in Figure 3: (a) GFF; (b) SR; (c) DWT; (d) CNN; (e) MWGF; (f) MST-SR; (g) NSST; (h) GAN; (i) the proposed method.
Entropy 23 00410 g005
Figure 6. The fused images of Group 3 in Figure 3: (a) GFF; (b) SR; (c) DWT; (d) CNN; (e) MWGF; (f) MST-SR; (g) NSST; (h) GAN; (i) the proposed method.
Figure 6. The fused images of Group 3 in Figure 3: (a) GFF; (b) SR; (c) DWT; (d) CNN; (e) MWGF; (f) MST-SR; (g) NSST; (h) GAN; (i) the proposed method.
Entropy 23 00410 g006aEntropy 23 00410 g006b
Figure 7. The fused images of Group 4 in Figure 3: (a) GFF; (b) SR; (c) DWT; (d) CNN; (e) MWGF; (f) MST-SR; (g) NSST; (h) GAN; (i) the proposed method.
Figure 7. The fused images of Group 4 in Figure 3: (a) GFF; (b) SR; (c) DWT; (d) CNN; (e) MWGF; (f) MST-SR; (g) NSST; (h) GAN; (i) the proposed method.
Entropy 23 00410 g007aEntropy 23 00410 g007b
Figure 8. The valuable metrics of the fused images in Figure 3: (a) entropy (EN); (b) average gradient (AVG); (c) spatial frequency (SF); (d) mutual information (MI).
Figure 8. The valuable metrics of the fused images in Figure 3: (a) entropy (EN); (b) average gradient (AVG); (c) spatial frequency (SF); (d) mutual information (MI).
Entropy 23 00410 g008
Figure 9. Oslo city: (a) SAR image; (b) optical image.
Figure 9. Oslo city: (a) SAR image; (b) optical image.
Entropy 23 00410 g009
Figure 10. The fused images of Figure 8: (a) GFF; (b) SR; (c) DWT; (d) CNN; (e) MWGF; (f) MST-SR; (g) NSST; (h) GAN; (i) the proposed method.
Figure 10. The fused images of Figure 8: (a) GFF; (b) SR; (c) DWT; (d) CNN; (e) MWGF; (f) MST-SR; (g) NSST; (h) GAN; (i) the proposed method.
Entropy 23 00410 g010
Table 1. The details of the generator and discriminator.
Table 1. The details of the generator and discriminator.
LayerFilterNormalizationActivation
G EncoderEn_15*5 Conv (n64)BNLeaky ReLU
En_23*3 Conv (n128)BNLeaky ReLU
En_3-En_53*3 Conv (n256)BNLeaky ReLU
DecoderDe_13*3 Conv (n256)BNLeaky ReLU
De_23*3 Conv (n128)BNLeaky ReLU
De_35*5 Conv (n1)-Sigmoid
D D_13*3 Conv (n64)BNLeaky ReLU
D_23*3 Conv (n128)BNLeaky ReLU
D_33*3 Conv (n256)BNLeaky ReLU
D_43*3 Conv (n1)-Sigmoid
Table 2. Objective indicators of generalization on 10 test images from the SEN1–2 datasets.
Table 2. Objective indicators of generalization on 10 test images from the SEN1–2 datasets.
ENAVGSFMI
GFF7.2607 ± 0.013210.0256 ± 0.063624.6531 ± 0.25014.8751 ± 0.0082
SR7.2251 ± 0.01619.6989 ± 0.108723.6590 ± 0.42916.8254 ± 0.0147
DWT7.2426 ± 0.011410.1984 ± 0.033225.2251 ± 0.26196.3567 ± 0.0258
CNN7.2675 ± 0.05039.9878 ± 0.140123.2157 ± 0.40694.6531 ± 0.0074
MWGF7.2475 ± 0.033510.1538 ± 0.040825.3621 ± 0.25856.4256 ± 0.0361
MST-SR7.2659 ± 0.035410.1596 ± 0.039525.0697 ± 0.26046.3751 ± 0.0292
NSST7.3105 ± 0.12069.5635 ± 0.153723.2758 ± 0.43134.9253 ± 0.0102
GAN7.2159 ± 0.080210.3756 ± 0.081925.5327 ± 0.36084.7754 ± 0.0146
Proposed7.4225 ± 0.020510.8597 ± 0.061126.4568 ± 0.25037.5754 ± 0.0319
Table 3. The valuable metrics of the fused images in Figure 10.
Table 3. The valuable metrics of the fused images in Figure 10.
ENAVGSFMITime(s)
GFF7.168410.994625.88431.12250.864875
SR7.363112.301630.36223.441677.549845
DWT7.344912.186630.23163.804430.458764
CNN7.356613.438632.20771.4334141.987512
MWGF7.454312.696331.20146.31483.648574
MST-SR7.456112.756031.37766.683171.457981
NSST7.429313.221831.70402.10164.987545
GAN7.381513.893432.03521.542858.145457
Proposed7.469414.769932.45437.620653.125794
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fang, J.; Ma, X.; Wang, J.; Qin, K.; Hu, S.; Zhao, Y. A Noisy SAR Image Fusion Method Based on NLM and GAN. Entropy 2021, 23, 410. https://doi.org/10.3390/e23040410

AMA Style

Fang J, Ma X, Wang J, Qin K, Hu S, Zhao Y. A Noisy SAR Image Fusion Method Based on NLM and GAN. Entropy. 2021; 23(4):410. https://doi.org/10.3390/e23040410

Chicago/Turabian Style

Fang, Jing, Xiaole Ma, Jingjing Wang, Kai Qin, Shaohai Hu, and Yuefeng Zhao. 2021. "A Noisy SAR Image Fusion Method Based on NLM and GAN" Entropy 23, no. 4: 410. https://doi.org/10.3390/e23040410

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop