Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Research on Public Service Request Text Classification Based on BERT-BiLSTM-CNN Feature Fusion
Previous Article in Journal
Analyzing the Impact of a Structured LLM Workshop in Different Education Levels
Previous Article in Special Issue
Haze-Aware Attention Network for Single-Image Dehazing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Learning Enriched Features Mechanism Using Deep Convolutional Neural Network for Image Denoising and Super-Resolution

1
University Institute of Information Technology, PMAS Arid Agriculture University Rawalpindi, Rawalpindi 46000, Pakistan
2
Department of Computer Science & Information Technology, University of Mianwali, Mianwali 42200, Pakistan
3
Department of Computer Science, Aberystwyth University, Penglais, Aberystwyth SY23 3DB, UK
4
Department of Physics, Physical Sciences Building, Aberystwyth University, Aberystwyth SY23 3BZ, UK
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2024, 14(14), 6281; https://doi.org/10.3390/app14146281
Submission received: 14 June 2024 / Revised: 15 July 2024 / Accepted: 15 July 2024 / Published: 18 July 2024
(This article belongs to the Special Issue Advances in Image Enhancement and Restoration Technology)

Abstract

:
Image denoising and super-resolution play vital roles in imaging systems, greatly reducing the preprocessing cost of many AI techniques for object detection, segmentation, and tracking. Various advancements have been accomplished in this field, but progress is still needed. In this paper, we have proposed a novel technique named the Enhanced Learning Enriched Features (ELEF) mechanism using a deep convolutional neural network, which makes significant improvements to existing techniques. ELEF consists of two major processes: (1) Denoising, which removes the noise from images; and (2) Super-resolution, which improves the clarity and details of images. Features are learned through deep CNN and not through traditional algorithms so that we can better refine and enhance images. To effectively capture features, the network architecture adopted Dual Attention Units (DUs), which align with the Multi-Scale Residual Block (MSRB) for robust feature extraction, working sidewise with the feature-matching Selective Kernel Extraction (SKF). In addition, resolution mismatching cases are processed in detail to produce high-quality images. The effectiveness of the ELEF model is highlighted by the performance metrics, achieving a Peak Signal-to-Noise Ratio (PSNR) of 42.99 and a Structural Similarity Index (SSIM) of 0.9889, which indicates the ability to carry out the desired high-quality image restoration and enhancement.

1. Introduction

Images are being used across various domains of today’s digital world. Examples of those areas include photography, digital entertainment, computer vision, remote sensing, medical diagnostics, microscopy, space science, and surveillance. Unfortunately, images tend to suffer from degradation during their formation and transmission processes. Such degradations come in many forms, such as noise, blur, intensity non-uniformity, missing pixels because of electronic or sensor failures, and interference due to neighboring electronic devices. These degradations will not only influence the visual quality of images but will also reduce their interpretability, and the effectiveness of image analysis and image processing algorithms. As part of this, restoring and enhancing degraded images is a very important and demanding area that aims to reduce the influence of various degradations and enhance the quality and interpretability of such images.
The purpose of enhancement and restoration techniques is to increase the visual appearance and perceptual quality of images, thus increasing their effectiveness for various desired applications such as image recognition [1], visual object detection [2], and semantic segmentation [3]. Enhancement techniques are assigned to improve particular visual qualities such as contrast, sharpness, or color balance. Image restoration techniques, on the other hand, focus on particular kinds of degradation in the image.
Image restoration techniques include denoising to remove noise, deblurring processes to recover sharpness, and inpainting to recover missing regions. Noisy images are very important to remove for interpretation and processing purposes. These types of images are inefficient for processing. Some fundamental unsupervised or supervised algorithms include face print-based identification, hidden identification [4], object detection purposes [5], and image segmentation.
Over time, significant developments have been made in image restoration and enhancement, powered by advancements in computer vision, machine learning, and signal processing [6]. Traditional approaches depend on algorithms that influence mathematical models and traditional rules. However, these methods often struggle with complex degradation patterns and require robust generalization across diverse image types.
In recent years, deep learning has changed the way images are processed and restored. Platforms like CNNs and all types of models that are part of deep learning, either encoder–decoder networks [7,8,9,10] or high-resolution models [11,12,13,14], are surprisingly good for image processing. A large amount of data in deep learning models and large-scale training makes it possible to learn complex mappings from the input (degraded image) to the output (pristine image) and to learn the task much better than traditional models, leading to more accurate and subjectively realistic image restoration and enhancement.
Our method provides a novel approach to image restoration and enhancement. The proposed approach addresses two specific problems in this area: denoising and super-resolution. Our main goal is to deeply explore the up-to-date and futuristic methods for image restoration and enhancement across the two specific challenges we have chosen. Through this process, our goal was to find unique supervised deep CNN techniques that will help improve image restoration and enhancement based on generated low-resolution images. Ultimately, we try to contribute by providing novel research into the development of image restoration and enhancement techniques while relying on aspects of supervised CNNs.
To start our discussions, we take up the topic of denoising a noisy image. Denoising is known to be a classically difficult problem since noise reduction tends to remove the fine details that are important for understanding; that is why classic methods of denoising either tend to provide a lot of noise along with details or tend to eradicate details along with noise. However, recent advances in deep learning-based techniques of denoising have revolutionized the domain of image noise reduction, where we use convolutional neural networks that help the model clear out as much noise as possible while preserving all the minute details of the image. These techniques have significantly increased the effectiveness of image denoising, but a lot of work still needs to be done in this domain.
In the next phase, we focus on the subfield of super-resolution. Super-resolution is a task that aims to restore image resolution beyond its original extent. In other words, this process “enhances” the quality of image resolution such that extra details are revealed, resulting in clearer visual appearances. We investigate the top-notch deep learning frameworks that are sketched specifically for super-resolution. These frameworks utilize neural networks to build up high-quality images from their low-quality equivalents, requiring intricate network designs and training schemes to improve image resolution while maintaining important structural attributes.
This approach aims to create a model that provides better visual results with more improved images and fewer computational complexities. In each resolution stage, the information has been exchanged hierarchically across all the scales. In contrast, traditional methods isolate each scale and process them in a top-down order, making this approach different from conventional methods. The processing of information exchange has been carried out via a kernel fusion process per stream. In addition, ELEF uses a self-attention mechanism after picking a useful set of kernels from each stream. The most significant part is the fusion of traits from varying receptive domains through a fusion block, keeping distinct features.

2. Related Work

The image restoration problem has been well-examined with the advancement in techniques over the past few years [15,16,17,18]. Several methods were introduced with various threats and trials in different restoration fields [14,19,20]. In the modern era, techniques including trainable neural networks replace conventional techniques [21,22,23,24], even skipping pre-assumption degradation procedures. The introduction of transformers provides new ways to approach the restoration domain. Some of them were originally designed for NLP tasks [25]. Vision Transformers break down the image into sequential patches, study their dependencies through relativity and capabilities, and illustrate images by cultivating the input data that entirely depend on self-attention [26]. Later, they are used for denoising, super-resolution, de-raining, and image colorization tasks [27,28,29]. The advancement in transformers that lessens the complications while developing sharper visions gives more precise results [30,31,32]. In addition, low-rank factorization and approximation approaches [33,34,35,36] ignite transformers yet lead to information loss, parameters, and task dependency.
Denoising. One crucial problem in image restoration is image denoising, which targets removing undesired noise while retaining important details. The traditional denoising approaches are based on filters like the median filter, the Wiener filter, and the wavelet transform, and on transforming coefficients and masking. Several patch-based methods [37,38,39] were also introduced, which resulted in redundancy in visuals. These types of techniques work in either the spatial or frequency domain. They are also built on simple assumptions about the noise (uniformly Gaussian distributed or multiplicative white noise) and/or the image (e.g., Gaussian distribution of DCT coefficients, piecewise smooth behavior) conditions. Deep learning-based methods have made huge progress in image denoising [40,41,42,43,44,45]. Convolutional Neural Networks (CNNs) have yielded impressive image denoising by effectively learning noise patterns (filters) and their handling. The latest DCNN (Deep Convolutional Neural Network) obtained the best performance in a wide range of denoising scenarios.
Super-Resolution. Super-resolution is the process of generating corresponding high-resolution (HR) images or videos from one or more low-resolution (LR) observations. The traditional methods are mainly devoted to statistical models or interpolation. However, to generate HR images with a natural appearance, including sharp edges and fine textures, many of them were based on sampling theory [46], edge-guided interpolation [47], natural image generation [48], and sparse illustration [49]. Recently, the progress of learning techniques for classification and other categories has also been remarkable. There are super-resolution (SR) approaches based on deep learning, such as the super-resolution (SR) reconstruction based on a simple CNN (SRCNN) [11] model and the non-linear mapping of high-resolution (SR) reconstruction based on CNN models (EDSR) [39] that follow the improvement of the model results. Deep learning techniques help to learn more about the inherent characteristics of high-resolution images. Different data-based techniques have different design frameworks [50,51]. While traditional methods directly produce HR images from LR images [52,53,54,55], modern techniques introduce residual learning architecture [55] for processing high-frequency image information. Apart from that, dense connections [56,57], multi-branch learning [58,59], progressive reconstruction [60], generative adversarial networks (GANs) [13,50,61,62], and recursive learning are also some of the methods for performing super-resolution tasks. Fraunhofer Institute in Germany has developed methods based on recursive learning [63], Non-local Means Filtering based on Variational Models for SR, and exponential cross-diffusion. Based on trained dictionaries, conventional SR methods show better performance in image quality using the dictionary-based high-frequency energy for details of HR images, and simulation results prove that our estimator achieves more precise texture reconstruction in high resolution.
Denoising followed by super-resolution: Few research studies have been conducted on the resolution of noisy images. Singh et al. [64] performed the two tasks separately and then joined the noisy super-resolution task and denoised super-resolution task. Laghrib et al. [65] performed the combined task by combining the denoising task with a newly introduced filter-based algorithm for super-resolution. Hu et al. [66] performed the super-resolution task with a simultaneous denoising task derived from a multi-scale noise reduction method. These methods do not leverage neural networks and fail to fully utilize edge information constraints. Chen et al. [67] used GAN for super-resolution and denoising tasks. They utilized the residual network to directly map the image with the original noise map. However, this approach does not fully exploit the constraints of edge information.

3. Method

We first take a sequential approach, taking care of the most fundamental task of denoising, and super-resolution as a subsequent task. We address the first step of denoising, where we want to refine an image to remove noise. By getting rid of noise, it lets the super-resolution step build a sharp image based on a consistent and relatively noise-free foundation, in essence preserving character and improving visual form, as shown in Figure 1.
Our overall pipeline consists of a series of modules that progressively improve image quality at each stage. The Multi-Scale Residual Blocks (MSRBs) are used to extract and represent different features. The Selective Kernel Feature Fusion (SKF) strengthens feature representation by fusing salient features. The Dual Attention Units (DUs) with feature Channel Attention (CA) and Spatial Attention (SA) mechanisms are used to refine features and enable each reforming process to focus on their restoration. The Residual Resizing Modules smoothly increase or decrease resolution without losing important image features.
Building on this generalization, we propose a simple two-step restoration pipeline: we perform L1-principal component analysis (L1-PCA) denoising before any of the super-resolution tasks. We aim to obtain highly compelling reconstruction results that are visually pleasant and perceptually accurate for various image domains and noise types. By forging the latest technologies into our pipeline and through our careful consideration of various details, our study aims to approach the problem of image restoration beyond cutting-edge leading ways to new image restoration and enhancement methodologies.

3.1. Overall Pipeline

The overall pipeline of our proposed restoration method follows a sequential flow of operations to enhance and restore the input image progressively. The input image is first fed to a sequence of MSRB modules as shown in Figure 2, thereby capturing and learning multi-scale features at different levels of abstraction efficiently. By doing this, the model can seize low-level to high-level information and hence make the restoration process easier.
R e s t o r e d = R e s i d u a l I n p u t + R e s i z e ( I n p u t )
In this pipeline, we first process the original image with the restoration network R e s i d u a l ( I n p u t ) to obtain the residual image, and then we upscale the input data back to the recovered residual image’s size with the R e s i z e ( I n p u t ) operation. Finally, we add the resized input image to the restored residual image and obtain the restored output image as the output. The idea of this pipeline is that the restored image can well preserve the details and structure in the original input image and can also include the restoration enhancements that are provided by the network in the residual image. The pipeline tries to make the restoration result visually pleasing by effectively removing the degradation artifacts in the degraded image and enhancing the overall quality of the degraded image.

3.1.1. Residual Resizing Modules (RRMs)

To mitigate the potential noisy discrepancy between input and output images, we introduce the residual resizing modules, which can be viewed as patches of the image revised with respect to noisy appearance. They are useful for ensuring that the final restored image remains the same structure as the clean one and does not delete important supporting details.

3.1.2. Multi-Scale Residual Block (MSRB)

The MSRB module is used as the basic component in our restoration network. It is composed of several convolutional layers with a residual connection, which is helpful to optimize the restoration through so-called residual learning. Different scales of convolutional filters are integrated into each MSRB to capture multi-scale features, and this improves the feature learning ability of the network in dealing with complex variations like texture change, edge change, and structure change inside the images.

3.1.3. Selective Kernel Feature Fusion (SKF)

We are further proposing a Selective Kernel Fusion (SKF) module which allows the network to select the features from all the convolutional features with different kernel sizes. This fusion gives the network the ability to capture effective local and global contextual information. By incorporating features from convolutional layers with different kernel sizes in addition to the mean, we further increase the feature representation capability of the network. As a result, the network can better restore details while inheriting more global context.
S K F = i = 1 N   ( j = 1 N W j W i F i )
SKF means the output feature map after the SKF module, F i represents the input feature maps of the different convolutional kernels, W i stands for the respective weight value corresponding to each activation map i, and N is the quantity of the activation maps. In the equation, ⊙ stands for the element-wise multiplication, and the normalization via the sum of the weights for each feature map in the denominator ensures the fusion operation. The SKF module selectively fuses the features of the different convolutional kernels to allow the network to effectively capture the local and global context information, which is beneficial for the network to better restore and enhance by adaptively weighting and combining the features to enrich the representation network potential.

3.1.4. Dual Attention Unit (DU)

The Dual Attention Unit (DU) module is intended to update the feature representations by introducing feature channel attention (CA) [68] and spatial attention (SA) [69] mechanisms. The channel attention focuses on modeling the dependencies between different channels, allowing the network to reinforce more useful features and alleviate less informative ones. The spatial attention focuses on modeling the correlation of different spatial locations in the feature maps, enabling the network to selectively emphasize the important regions. Through the cooperation of CA and SA, the DU module can improve the network’s discriminant ability and achieve accurate restoration effectively.

Channel Attention (CA)

The CA [68] module can calculate channel-focused attention weights to obtain the informative features and denoise the noisy or irrelevant information by focusing on global statistics. Mathematically, the channel attention mechanism can be represented as follows:
C A ( x ) = σ ( W 2 δ ( W 1 x ) )
where x represents the input feature map, δ denotes the ReLU activation function, W1 and W2 are learnable weights, and σ represents the sigmoid function.

Spatial Attention (SA)

The SA mechanism extracts a spatial attention map, which identifies locally important spatial regions, from the feature maps and emphasizes the interactions among the neighboring spatial locations. The SA [69] mechanism models spatial dependencies. Mathematically, the spatial attention mechanism can be represented as follows:
S A ( x ) = σ ( W 3 δ ( W 4 x ) )
where x represents the input feature map, δ denotes the ReLU activation function, W3 and W4 are learnable weights, and σ represents the sigmoid function. Our model uses adjustable weights W1, W2, W3 and W4, in the range [0.5–1.5]. Any value below or above significantly affects the model’s performance on denoising and super-resolution tasks. The best performance was observed when all weights were set to 1.0, indicating that balanced attention mechanisms work well for the said tasks.

4. Experiments and Results

To assess the performance of the proposed method for IR, restoration, and enhancement using learning enriched features, experiments have been conducted separately for denoising and super-resolution tasks. Experiments are performed on the publicly available Diverse IR Image Datasets SIDD [70] and DND [71], which contain real-life captured degraded IR images under different types of environments.
Before enhancing the image, we will denoise the noisy images first. This is because noise artifacts always cover many details of the image, and noise may also cause some non-existing artifacts to appear in the image. Hence, we must filter out the noise first. By doing this, we can keep only the important details and then enhance the important details in the low-resolution image to a higher resolution. The dataset comprises a large number of real-world images with various forms of degradations. The dataset’s images are from noisy environments, as in real applications. The dataset exhibits various noise-level images from different scenes and objects. Thus, the dataset can represent almost all the possible noise level conditions in the real world.

4.1. Datasets

The Smartphone Image Denoising Dataset (SIDD) [70] is a benchmarked dataset with real-world noisy images under diverse lighting and ISO conditions, photographed with smartphones. It contains a variety of noise levels in the images. We use a total of 1600 pairs, of which 320 image pairs are used for training and 1280 are used for validation. The datasets are prepared following a series of processing steps to handle camera shift alignment, exposure time adjustment, and intensity scaling. Some sample images of DND and SIDD dataset are shown in Figure 3.
The Darmstadt Noise Dataset (DND) [71] is a benchmark dataset including 50 pairs, each consisting of real noisy and ground-truth data captured with various consumer-grade cameras. Taking into consideration the ISO, noisy images and reference images are taken with higher and base levels, respectively. As high-resolution images are used, it contains 20 extracted crops of size 512 × 512 per image, resulting in a total of 1000 patches. All of these are used for testing (DND lacks training or validation sets). The ground-truth noise-free images are not publicly released, so an online server is used to provide images for quantitative measures.

4.2. Training Dataset Setup

To train our enhancement model effectively, we require a dataset that includes pairs of low-quality images and their corresponding high-quality reference images. These pairs of images act as the ideal output images, used as a benchmark for the model’s performance. So, this is the ground truth that the network can use to learn the enhancement mapping accurately. The dataset must include a varied range of image variations, like different levels of degradation, different scenes, and different objects. This is important because the scenarios may change from image to image. The category, scene, and objects in the scene may change for every image. So, to make our model generalize well and perform well on all types of images, we require these parameters in our dataset. Our goal is to make our model work well on all types of real-world images, so if we train our model using such a comprehensive dataset, then the network will have learned very robust enhancement techniques that can be applied to all real-world images.

4.3. Experimental Setup

Our deep convolutional neural network model was implemented using the PyTorch library in Google Collaboratory (collab), which is well-suited to Deep Learning frameworks, and running it on a high-performance computing cluster with NVIDIA GPUs. The dataset was randomly split into training, validation, and testing sets with a ratio of 80:10:10, respectively. Table 1 shows the basic characteristics of the datasets used for comparison purposes. The Dual Attention Unit (DU), Residual Residing Modules (RRMs), and Selective Kernel Feature Fusion (SKF) are the same as in MIRNet [72].

4.4. Performance Measures

To quantitatively assess the effectiveness of our method, we have used a couple of evaluation metrics, including Peak Signal-to-Noise Ratio (PSNR) [70] and Structural Similarity Index Matrix (SSIM) [71]. These metrics provide a clue to the quality, similarity, and error between the restored/enhanced images and the ground truth high-quality images.

4.5. Analysis with Baseline Methods

We compared our proposed method with several state-of-the-art and well-established image restoration and enhancement techniques, including conventional approaches and deep learning frameworks, on standard benchmark datasets SIDD [70] and DND [71]. Figure 4 and Figure 5 show denoising and super-resolution results in terms of PSNR and SSIM on the SIDD benchmark dataset for MIRNet [72], RIDNet [73], CBDNet [74], and our proposed ELEF mechanism. It is clear from the results in Figure 4 and Figure 5 that CBDNet still has issues with edge preservation and detail preservation, whereas RIDNet and MIRNet preserve edges but lack detail preservation. Our proposed model outperforms other competing mechanisms in edge preservation as well as other detail preservation, with better PSNR and SSIM values.
Figure 6 displays denoising and super-resolution results in terms of PSNR and SSIM on the DND benchmark dataset for MIRNet [72], RIDNet [73], CBDNet [74], VDN [75] and our proposed ELEF mechanism. It can be observed from the results in Figure 6 that MIRNet, RIDNet, VDN, and CBDNet have some blurry effects and smooth out some parts of edges, whereas our proposed model gives excellent results in terms of detail preservation and edge preservation without any deterioration. Furthermore, the proposed filter gives excellent PSNR and SSIM values and outperforms other models.
The proposed model is evaluated and tested quantitatively in terms of PSNR and SSIM for other models as well in Table 2 and Table 3 using the benchmark datasets mentioned above. It is evident from Table 2 and Table 3 that the proposed model gives much better PSNR and SSIM values when compared with other state-of-the-art models.

4.6. Qualitative Results

The proposed method also achieves significantly better restoration and enhancement results in terms of visual effects. To assess the quality of our results, we performed a survey on our university campus and showed the results to 75 participants (students, faculty, and staff). Most of the participants (72) gave satisfactory remarks, except for a few (3) who did not find any differences, resulting in a 96% success rate. It can be clearly seen from the results shown in Figure 4, Figure 5 and Figure 6 that the noise is significantly reduced, the details are enhanced, and the images are sharpened when compared with the input images.

5. Ablation Studies

We investigated the impact of our architectural components and design choices on final performance through a series of ablation experiments conducted on image denoising and super-resolution tasks, as shown in Table 4. Table 4 highlights that the absence of skip connections leads to the most significant decline in performance. Without these connections, the network faces convergence issues, resulting in higher training errors and lower PSNR. Additionally, the Selective Kernel Feature Fusion (SKF) mechanism, which facilitates information exchange among parallel convolution streams, proves advantageous and boosts performance. Likewise, the Dynamic Attention Units (DUs) contribute positively to the overall image quality.
Table 5 shows feature combinations with summation (SUM), concatenation (CAT), and Selective Kernel Feature Fusion (SKF). The proposed SKF is better than SUM and CAT, utilizing ∼6 times fewer parameters than CAT and generating better PSNR results. Specifically, SKF effectively enhances feature representation by adaptively selecting and combining informative features from parallel convolution streams. Additionally, a significant reduction in parameters highlights the efficiency and effectiveness of SKF in improving overall model performance.
In Table 6, we perform an experiment on the RealSR [8] dataset. For the denoising task, it has not shown much difference, as the dataset has no noise. Still, the super-resolution task performs better visually and gets higher PSNR and SSIM values, which shows the significance of the method in real-world scenarios. Although it is a computationally resource-oriented task, as the process of denoising has been performed on a denoised dataset, the denoising mechanism still improves the image quality.
Furthermore, in our ablation study, as shown in Figure 7, we found that each module within the ELEF model significantly contributes to its overall performance in image restoration and enhancement tasks. When the MSRB (Multi-Scale Residual Block) was removed, our model struggled to capture fine details, resulting in slightly softer restored images, as shown in Figure 7. Excluding the SKF (Selective Kernel Fusion) module affected the model’s ability to integrate information across the image, leading to less natural-looking enhancements. Without the DU (Dual Attention Unit), the model had difficulty focusing on important image features, resulting in slightly noisier or less sharp restorations. Lastly, removing the RRM (Residual Refinement Module) meant that the model could not refine images as effectively, leaving some minor imperfections. These findings highlight the importance of each module (MSRB, SKF, DU, RRM) in enhancing the ELEF model’s performance and guiding future improvements to achieve better restoration quality in various applications.
In addition, we have analyzed our results on different resolutions: ×2, ×3, and ×4, as shown in Figure 8 and Figure 9. It can be seen from the visual results that with the increase in resolution, the quality of the visual results deteriorated. So, it’s better to use the proposed mechanism below ×3 resolution for image processing and computer vision tasks.

6. Conclusions

The experimental results can prove the superiority of the proposed algorithm in various noise and blurring removal tasks, which can be used in image restoration and enhancement. It takes into account the introduction of learning enriched features, multi-scale residual blocks, selective kernel feature fusion, dual attention units, and residual size variation modules in deep learning for dealing with different types of imaging noise and blurring degradation. The results are visually attractive and perceptually moderately accurate. There has been a breakthrough in the field of video resolution and damage, particularly in noise removal, followed by video resolution and the use of deep learning technologies. The algorithm developed in this paper makes full use of deep learning technology and other advanced technologies such as learning enriched features in this field. Evaluated against competitive methods, the algorithm has outstanding advantages in terms of image output quality and performance. Through various tests, the proposed method can adapt to real scenes, with high definition and rich details and colors. In light of these factors, imaging systems, monitoring, remote sensing, and other scenarios can clearly and accurately display the target image, which has great advantages for precision analysis and judgment.

Author Contributions

Conceptualization, I.W. and M.H.; Data curation, M.A. and S.F.J.; Formal analysis, E.R. and S.F.J.; Funding acquisition, M.A.; Methodology, I.W. and M.H.; Supervision, M.H.; Validation, E.R.; Visualization, R.M.Y.; Writing—original draft, I.W. and M.H.; Writing—review and editing, R.B., R.M.Y. and M.W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 4690–4699. [Google Scholar]
  2. Han, W.; Chang, S.; Liu, D.; Yu, M.; Witbrock, M.; Huang, T.S. Image super-resolution via dual-state recurrent networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  3. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  4. Kim, Y.; Soh, J.W.; Park, G.Y.; Cho, N.I. Transfer learning from synthetic to real-noise denoising with adaptive instance normalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 3482–3492. [Google Scholar]
  5. Kim, J.; Kwon Lee, J.; Mu Lee, K. Deeply-recursive convolutional network for image super-resolution. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  6. Wali, A.; Naseer, A.; Tamoor, M.; Gilani, S.A.M. Recent Progress in Digital Image Restoration Techniques: A Review. Digit. Signal Process. 2023, 141, 104187. [Google Scholar] [CrossRef]
  7. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015, Proceedings, Part III 18; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  8. Kupyn, O.; Martyniuk, T.; Wu, J.; Wang, Z. Deblurgan-v2: Deblurring (orders-of-magnitude) faster and better. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8878–8887. [Google Scholar]
  9. Chen, C.; Chen, Q.; Xu, J.; Koltun, V. Learning to see in the dark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  10. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  11. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [PubMed]
  12. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
  13. Sajjadi, M.S.; Scholkopf, B.; Hirsch, M. Enhancement: Single image super-resolution through automated texture synthesis. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  14. Ignatov, A.; Kobyshev, N.; Timofte, R.; Vanhoey, K.; Van Gool, L. Dslr-quality photos on mobile devices with deep convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3277–3285. [Google Scholar]
  15. Fattal, R. Image upsampling via imposed edge statistics. In Proceedings of the ACM Special Interest Group on Computer Graphics and Interactive Techniques Conference (SIGGRAPH), San Diego, CA, USA, 5–9 August 2007; p. 95-es. [Google Scholar]
  16. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [PubMed]
  17. Kopf, J.; Neubert, B.; Chen, B.; Cohen, M.; Cohen-Or, D.; Deussen, O.; Uyttendaele, M.; Lischinski, D. Deep photo: Model-based photograph enhancement and viewing. ACM Trans. Graph. (TOG) 2008, 27, 1–10. [Google Scholar] [CrossRef]
  18. Michaeli, T.; Irani, M. Nonparametric blind super-resolution. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013; pp. 945–952. [Google Scholar]
  19. Abdelhamed, A.; Timofte, R.; Brown, M.S. Ntire 2019 challenge on real image denoising: Methods and results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  20. Li, Y.; Zhang, Y.; Timofte, R.; Van Gool, L.; Tu, Z.; Du, K.; Wang, H.; Chen, H.; Li, W.; Wang, X.; et al. NTIRE 2023 challenge on image denoising: Methods and results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 1904–1920. [Google Scholar]
  21. Chen, L.; Chu, X.; Zhang, X.; Sun, J. Simple baselines for image restoration. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 17–33. [Google Scholar]
  22. Zhang, Y.; Li, D.; Shi, X.; He, D.; Song, K.; Wang, X.; Qin, H.; Li, H. Kbnet: Kernel basis network for image restoration. arXiv 2023, arXiv:2303.02881. [Google Scholar]
  23. Chen, L.; Lu, X.; Zhang, J.; Chu, X.; Chen, C. Hinet: Half instance normalization network for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 182–192. [Google Scholar]
  24. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Multi-stage progressive image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14821–14831. [Google Scholar]
  25. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  26. Ghasemabadi, A.; Salameh, M.; Janjua, M.K.; Zhou, C.; Sun, F.; Niu, D. CascadedGaze: Efficiency in Global Context Extraction for Image Restoration. arXiv 2024, arXiv:2401.15235. [Google Scholar]
  27. Efros, A.A.; Leung, T.K. Texture synthesis by non-parametric sampling. In Proceedings of the 7th IEEE International Conference on Computer Vision (ICCV), Corfu, Greece, 20–25 September 1999. [Google Scholar]
  28. Freedman, G.; Fattal, R. Image and video upscaling from local self-examples. ACM Trans. Graph. (ToG) 2011, 30, 1–11. [Google Scholar] [CrossRef]
  29. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Learning enriched features for real image restoration and enhancement. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part XXV 16; Springer International Publishing: Cham, Switzerland, 2020; pp. 492–511. [Google Scholar]
  30. Yaroslavsky, L.P. Local adaptive image restoration and enhancement with the use of DFT and DCT in a running window. In Proceedings of the Wavelet Applications in Signal and Image Processing IV, Denver, CO, USA, 6–9 August 1996. [Google Scholar]
  31. Donoho, D.L. De-noising by soft-thresholding. Trans. on information theory. IEEE Trans. Inf. Theory 1995, 41, 613–627. [Google Scholar] [CrossRef]
  32. Simoncelli, E.P.; Adelson, E.H. Noise removal via Bayesian wavelet coring. In Proceedings of the International Conference on Image Processing (ICIP), Lausanne, Swittzerland, 16–19 September 1996. [Google Scholar]
  33. Smith, S.M.; Brady, J.M. SUSANa new approach to low level image processing. Int. J. Comput. Vis. 1997, 23, 45–78. [Google Scholar] [CrossRef]
  34. Tomasi, C.; Manduchi, R. Bilateral filtering for gray and color images. In Proceedings of the 6th International Conference on Computer Vision (ICCV-98), Bombay, India, 4–7 January 1998. [Google Scholar]
  35. Perona, P.; Malik, J. Scale-space and edge detection using anisotropic diffusion. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 629–639. [Google Scholar] [CrossRef]
  36. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  37. Dong, W.; Shi, G.; Li, X. Nonlocal image restoration with bilateral variance estimation: A low-rank approach. IEEE Trans. Image Process. 2012, 2, 700–711. [Google Scholar] [CrossRef]
  38. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  39. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the 12th International Conference on Computer Vision Workshops (ICCV), Kyoto, Japan, 29 September–2 October 2009. [Google Scholar]
  40. Hedjam, R.; Moghaddam, R.F.; Cheriet, M. Markovian clustering for the non-local means image denoising. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 3877–3880. [Google Scholar]
  41. Brooks, T.; Mildenhall, B.; Xue, T.; Chen, J.; Sharlet, D.; Barron, J.T. Unprocessing images for learned raw denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  42. Gharbi, M.; Chaurasia, G.; Paris, S.; Durand, F. Deep joint demosaicking and denoising. ACM Trans. Graph. (TOG) 2016, 35, 1–12. [Google Scholar] [CrossRef]
  43. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward convolutional blind denoising of real photographs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 1712–1722. [Google Scholar]
  44. Plötz, T.; Roth, S. Neural nearest neighbors networks. Adv. Neural Inf. Process. Syst. 2018, 31, 1095–1106. [Google Scholar]
  45. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed]
  46. Irani, M.; Peleg, S. Improving resolution by image registration. CVGIP Graph. Models Image Process. 1991, 53, 231–239. [Google Scholar] [CrossRef]
  47. Zhang, L.; Wu, X. An edge-guided image interpolation algorithm via directional filtering and data fusion. IEEE Trans. Image Process. 2006, 15, 2226–2238. [Google Scholar] [CrossRef] [PubMed]
  48. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image super-resolution via sparse. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef]
  49. Xiong, Z.; Sun, X.; Wu, F. Robust web image/video super-resolution. IEEE Trans. Image Process. 2010, 19, 2017–2028. [Google Scholar] [CrossRef]
  50. Wang, Z.; Chen, J.; Hoi, S.C. Deep learning for image super-resolution: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3365–3387. [Google Scholar] [CrossRef] [PubMed]
  51. Anwar, S.; Khan, S.; Barnes, N. A deep journey into super-resolution: A survey. arXiv 2019, arXiv:1904.07523. [Google Scholar] [CrossRef]
  52. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for. In Proceedings of the 13th European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  53. Kim, J.; Kwon Lee, J.; Mu Lee, K. Accurate image super-resolution using very. In Proceedings of the International Conference on Computer Vision (ICCV), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  54. Tai, Y.; Yang, J.; Liu, X.; Xu, C. Memnet: A persistent memory network for. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  55. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  56. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. ES-RGAN: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  57. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  58. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  59. Dahl, R.; Norouzi, M.; Shlens, J. Pixel recursive super resolution. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  60. Wang, Z.; Liu, D.; Yang, J.; Han, W.; Huang, T. Deep networks for image super-resolution with sparse prior. In Proceedings of the International Conference on Computer Vision (ICCV 2015), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  61. Park, S.J.; Son, H.; Cho, S.; Hong, K.S.; Lee, S. Srfeat: Single image super-resolution with feature discrimination. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  62. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  63. Nah, S.; Kim, T.H.; Lee, K.M. Deep Multi-Scale Convolutional Neural Network for Dynamic Scene Deblurring. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 257–265. [Google Scholar]
  64. Singh, A.; Porikli, F.; Ahuja, N. Super-resolving noisy images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2846–2853. [Google Scholar]
  65. Laghrib, A.; Ezzaki, M.; El Rhabi, M.; Hakim, A.; Monasse, P.; Raghay, S. Simultaneous deconvolution and denoising using a second order variational approach applied to image super resolution. Comput. Vis. Image Underst. 2018, 168, 50–63. [Google Scholar] [CrossRef]
  66. Hu, J.; Wu, X.; Zhou, J. Noise robust single image super-resolution using a multiscale image pyramid. Signal Process. 2018, 148, 157–171. [Google Scholar] [CrossRef]
  67. Chen, L.; Dan, W.; Cao, L.; Wang, C.; Li, J. Joint denoising and super-resolution via generative adversarial training. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2753–2758. [Google Scholar]
  68. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  69. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  70. Abdulhamed, A.; Lin, S.; Brown, M.S. A high-quality denoising dataset for smartphone cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1692–1700. [Google Scholar]
  71. Plotz, T.; Roth, S. Benchmarking denoising algorithms with real photographs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1586–1595. [Google Scholar]
  72. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Learning enriched features for fast image restoration and enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1934–1948. [Google Scholar] [CrossRef]
  73. Anwar, S.; Barnes, N. Real image denoising with feature attention. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3155–3164. [Google Scholar]
  74. Zhang Yue, Z.; Yong, H.; Zhao, Q.; Meng, D.; Zhang, L. Variational denoising network: Toward blind noise modeling and removal. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
  75. Burger, H.C.; Schuler, C.J.; Harmeling, S. Image denoising: Can plain neural networks compete with BM3D. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2392–2399. [Google Scholar]
  76. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  77. Mou, C.; Zhang, J.; Wu, Z. Dynamic attentive graph learning for image restoration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 4328–4337. [Google Scholar]
  78. Chang, M.; Li, Q.; Feng, H.; Xu, Z. Spatial-adaptive network for single image denoising. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part XXX 16; Springer International Publishing: Cham, Switzerland, 2020; pp. 171–187. [Google Scholar]
  79. Ren, C.; He, X.; Wang, C.; Zhao, Z. Adaptive consistency prior based deep network for image denoising. In Proceedings of the IEEE/CVF Conference on Computer vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 8596–8606. [Google Scholar]
  80. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Cycleisp: Real image restoration via improved data synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2696–2705. [Google Scholar]
  81. Cai, J.; Zeng, H.; Yong, H.; Cao, Z.; Zhang, L. Toward real-world single image super-resolution: A new benchmark and a new model. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  82. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR 2019), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  83. Tang, Y.; Han, K.; Guo, J.; Xu, C.; Xu, C.; Wang, Y. GhostNetv2: Enhance cheap operation with long-range attention. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; Volume 35, pp. 9969–9982. [Google Scholar]
  84. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
Figure 1. Flow diagram.
Figure 1. Flow diagram.
Applsci 14 06281 g001
Figure 2. Proposed model diagram.
Figure 2. Proposed model diagram.
Applsci 14 06281 g002
Figure 3. Sample images from SIDD [70] and DND [71] datasets.
Figure 3. Sample images from SIDD [70] and DND [71] datasets.
Applsci 14 06281 g003
Figure 4. Denoising and super-resolution results on the SIDD benchmark dataset.
Figure 4. Denoising and super-resolution results on the SIDD benchmark dataset.
Applsci 14 06281 g004aApplsci 14 06281 g004b
Figure 5. Proposed model results on the SIDD benchmark dataset.
Figure 5. Proposed model results on the SIDD benchmark dataset.
Applsci 14 06281 g005
Figure 6. Results on the DND benchmark dataset.
Figure 6. Results on the DND benchmark dataset.
Applsci 14 06281 g006aApplsci 14 06281 g006b
Figure 7. Performance analysis of ELEF without MSRB, SKF, DU, and RRM modules.
Figure 7. Performance analysis of ELEF without MSRB, SKF, DU, and RRM modules.
Applsci 14 06281 g007
Figure 8. Results for different resolutions on the SIDD dataset.
Figure 8. Results for different resolutions on the SIDD dataset.
Applsci 14 06281 g008
Figure 9. Results for different resolutions on the DND dataset.
Figure 9. Results for different resolutions on the DND dataset.
Applsci 14 06281 g009
Table 1. Components of benchmark datasets.
Table 1. Components of benchmark datasets.
ComponentsSSIM DatasetDND Dataset
SourceSmartphone camerasConsumer cameras
Number of Original Images16,000150
Noise LevelHighRelatively low
Data Provided320 image pairs (training), 1280 image pairs (validation)1000 image patches BREAK (512 × 512) for testing
Ground TruthAvailableAvailable
Table 2. Comparison on the SIDD [2] dataset. Highlights that methods are arranged on the bases of PSNR in ascending order.
Table 2. Comparison on the SIDD [2] dataset. Highlights that methods are arranged on the bases of PSNR in ascending order.
METHODSPSNR ↑SSIM ↑
DnCNN [76]23.620.581
MLP [77]24.720.643
BM3D [78]25.630.682
CBDNet [74]30.740.802
RIDNet [73]38.720.951
DAGL [79]38.920.951
VDN [75]39.230.952
SADNet [80]39.410.954
DeamNet [81]39.420.951
CycleISP [82]39.520.953
MIRNet-v2 [72]39.830.951
ELEF (Ours)42.990.9889
Table 3. Comparison on the DND [53] dataset. Highlights that methods are arranged on the bases of PSNR in ascending order.
Table 3. Comparison on the DND [53] dataset. Highlights that methods are arranged on the bases of PSNR in ascending order.
METHODSPSNR ↑SSIM ↑
DnCNN [76]32.410.791
MLP [77]34.220.834
BM3D [78]34.520.850
CBDNet [74]38.050.941
RIDNet [73]39.250.954
VDN [75]39.390.951
CycleISP [82]39.540.955
SADNet [80]39.580.954
DeamNet [81]39.640.952
DAGL [79]39.760.955
MIRNet-v2 [72]39.830.956
ELEF (ours)39.910.985
Table 4. Impact of different components of MSRBs.
Table 4. Impact of different components of MSRBs.
Components Presence of Components
Skip Connections
DU
SKF
PSNR (in dB)27.9030.5634.32
Table 5. Feature combinations.
Table 5. Feature combinations.
Methods SUMCATSKF
PSNR (in dB)30.7730.8834.32
Parameters012,2862048
Table 6. Denoising and super-resolution performed on the RealSR [8] dataset. Highlights that methods are arranged on the bases of PSNR in ascending order.
Table 6. Denoising and super-resolution performed on the RealSR [8] dataset. Highlights that methods are arranged on the bases of PSNR in ascending order.
ScalesBicubicRCAN [83]LP-KPN [84]MIRNet [72]ELEF (Ours) ↑
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
×232.620.90633.860.92133.910.92634.340.93437.160.939
×329.330.84130.410.86130.410.86731.150.88434.320.890
×427.980.80728.870.82528.930.83329.150.84431.260.851
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Waseem, I.; Habib, M.; Rehman, E.; Bibi, R.; Yousaf, R.M.; Aslam, M.; Jilani, S.F.; Younis, M.W. Enhanced Learning Enriched Features Mechanism Using Deep Convolutional Neural Network for Image Denoising and Super-Resolution. Appl. Sci. 2024, 14, 6281. https://doi.org/10.3390/app14146281

AMA Style

Waseem I, Habib M, Rehman E, Bibi R, Yousaf RM, Aslam M, Jilani SF, Younis MW. Enhanced Learning Enriched Features Mechanism Using Deep Convolutional Neural Network for Image Denoising and Super-Resolution. Applied Sciences. 2024; 14(14):6281. https://doi.org/10.3390/app14146281

Chicago/Turabian Style

Waseem, Iqra, Muhammad Habib, Eid Rehman, Ruqia Bibi, Rehan Mehmood Yousaf, Muhammad Aslam, Syeda Fizzah Jilani, and Muhammad Waqar Younis. 2024. "Enhanced Learning Enriched Features Mechanism Using Deep Convolutional Neural Network for Image Denoising and Super-Resolution" Applied Sciences 14, no. 14: 6281. https://doi.org/10.3390/app14146281

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop