Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Analysis of the Scenarios of Use of an Innovative Technology for the Fast and Nondestructive Characterization of Viscoelastic Materials in the Tires Field
Previous Article in Journal
Evaluating User Perceptions of a Vibrotactile Feedback System in Trunk Stabilization Exercises: A Feasibility Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CSINet: A Cross-Scale Interaction Network for Lightweight Image Super-Resolution

1
School of Computer Science and Engineering, Macau University of Science and Technology, Macau 999078, China
2
School of Electronic Information, Dongguan Polytechnic, Dongguan 523109, China
3
School of Computer Science, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(4), 1135; https://doi.org/10.3390/s24041135
Submission received: 11 January 2024 / Revised: 26 January 2024 / Accepted: 7 February 2024 / Published: 9 February 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In recent years, advancements in deep Convolutional Neural Networks (CNNs) have brought about a paradigm shift in the realm of image super-resolution (SR). While augmenting the depth and breadth of CNNs can indeed enhance network performance, it often comes at the expense of heightened computational demands and greater memory usage, which can restrict practical deployment. To mitigate this challenge, we have incorporated a technique called factorized convolution and introduced the efficient Cross-Scale Interaction Block (CSIB). CSIB employs a dual-branch structure, with one branch extracting local features and the other capturing global features. Interaction operations take place in the middle of this dual-branch structure, facilitating the integration of cross-scale contextual information. To further refine the aggregated contextual information, we designed an Efficient Large Kernel Attention (ELKA) using large convolutional kernels and a gating mechanism. By stacking CSIBs, we have created a lightweight cross-scale interaction network for image super-resolution named “CSINet”. This innovative approach significantly reduces computational costs while maintaining performance, providing an efficient solution for practical applications. The experimental results convincingly demonstrate that our CSINet surpasses the majority of the state-of-the-art lightweight super-resolution techniques used on widely recognized benchmark datasets. Moreover, our smaller model, CSINet-S, shows an excellent performance record on lightweight super-resolution benchmarks with extremely low parameters and Multi-Adds (e.g., 33.82 dB@Set14 × 2 with only 248 K parameters).

1. Introduction

Single image super resolution (SR) is a low-level task in the field of computer vision that aims to reconstruct a high resolution from a corresponding low-resolution image (LR), which is widely used in many applications, such as mobile devices, surveillance systems, autonomous driving, medical imaging, etc. However, SR is an ill-posed problem, since an identical LR image may be degenerated from different HR images. Therefore, SR is still a challenging task in terms of how to efficiently visually reconstruct HR images from degraded LR images.
To address this issue, Dong et al. [1] proposed SRCNN, marking the first application of deep learning methods in the field of single image super-resolution (SR). Achieving significantly superior results compared to traditional methods, SRCNN utilizes a three-layer convolutional neural network. To mitigate the computational demands of SRCNN, Kim et al. [2] introduced the VDSR model, incorporating residual learning to deepen the network to 20 layers and achieve rapid convergence. Lim et al. [3] presented the EDSR model, simplifying the network structure by removing batch normalization layers (BN), enhancing the model’s representational capacity, and securing victory in the NTIRE2017 Super-Resolution Challenge. Zhang et al. [4] proposed a residual-in-residual structure, pushing the depth of the Convolutional Neural Network (CNN) to 100–400 layers, yielding remarkably high PSNR values on benchmark datasets, surpassing previous methods.
While increasing the depth and width of CNNs can significantly enhance the network performance, it also comes with a significant computational cost and memory overhead, which restricts its use in practical applications like mobile devices, robots, edge computing, etc. In order to solve the issue, it is essential to develop a network that is both lightweight and incredibly effective.
Several prior methods [5,6,7,8,9,10] have been proposed in order to achieve a better trade-off between SR performance and computational efficiency. However, there are issues such as a small receptive field, slow convergence speed, information loss, and the fact that all network structures are constructed by a single-scale convolution kernel.
Recent research indicates that efficient operator design is crucial for building efficient SR CNNs. Factorized convolution, as an efficient operator, can break down standard convolution operations into two smaller convolution operations, effectively reducing the computational complexity and the number of parameters while maintaining network performance. However, there has been no prior study on efficient SR algorithms based on factorized convolution. Therefore, in this paper, we adopt a factorized convolution approach and construct an efficient Cross-Scale Interaction Block (CSIB).
The design rationale of CSIB is as follows: Firstly, we employ a 1 × 1 convolution layer to decrease the number of parameters and expedite the training process, followed by splitting them into two separate branches. To alleviate the computational burden, we employ factorized convolution to factorize the standard 5 × 5 depth-wise convolution into 5 × 1 and 1 × 5 depth-wise convolutions. One branch makes use of the factorized depth-wise convolution to extract local fine-grained features, while the other branch utilizes factorized depth-wise dilated convolution to capture global coarse-grained features. To prevent grid artifacts, different CSIB modules employ varying dilation rates. It is important to note that, for the more effective integration of cross-scale contextual information, we conduct interaction operations at the end of the dual-branch structure. This design not only reduces the computational complexity but also adeptly merges the local details and global features in the image. To further refine the aggregated contextual information, we designed an Efficient Large Convolutional Kernel Attention (ELKA) using large convolutional kernels and a gating mechanism.
We built a Cross-Scale Interaction Network, named “CSINet”, by stacking CSIBs to extract multi-scale contextual information. This design balances the performance and computational efficiency, and our CSINet outperforms most lightweight SISR methods in terms of both the performance and computational complexity. Specific experimental results can be seen in Figure 1. This is a summary of the key contributions of our work:
  • We adopted a factorized convolution approach to design a Cross-Scale Interaction Block (CSIB). CSIBs employ a dual-branch structure to extract both local fine-grained features and global coarse-grained features. Furthermore, we utilize interaction operations at the end of the dual-branch structure, facilitating the integration of cross-scale contextual information;
  • We designed an Efficient Large Convolutional Kernel Attention (ELKA) with limited additional computation for refining and extracting features. Ablation studies validated the effectiveness of this attention module;
  • Comprehensive experiments on benchmark datasets show that our CSINet outperforms most state-of-the art lightweight SR methods.

2. Related Work

2.1. Lightweight Image SR

To improve the network speed while maintaining superior reconstruction results, several lightweight image super-resolution networks have been introduced [1,7,11,12,13,14]. These networks can be broadly categorized into three groups: network structure design, knowledge distillation, and pruning. In the network structure design methods, FSRCNN [1] is the first lightweight super-resolution model. It performs upsampling at the end of the network, significantly improving the processing speed, but the performance of image reconstruction still needs improvement. CARN [14] designs a cascaded residual module based on grouped convolution and adopts a mechanism of local and global cascading to fuse multi-layer features, thereby accelerating the model’s running speed. PAN [8] designs self-calibrated blocks with pixel attention and upsampling blocks, achieving competitive performance with only 272K parameters.
In knowledge distillation methods, IDN [15] uses 1 × 1 and 3 × 3 convolutions to construct an information distillation module, distilling the current feature map through channel separation, achieving real-time performance while maintaining reconstruction accuracy. Based on IDN, IMDN [11] introduces a multi-information distillation module that extracts a part of useful features each time and integrates the remaining features into the distillation step of the next stage. After completion, the features extracted in each step are connected together. Subsequently, RFDN [13] combines feature distillation connections and shallow residual blocks to construct a residual feature distillation block, achieving a better performance than IMDN with fewer parameters.
In pruning methods, SCCVLAB [16] uses a fine-grained channel pruning strategy to address image super-resolution problems, achieving satisfactory results. SMSR [7] prunes redundant computations by learning spatial and channel masks, achieving a better performance with an improved inference efficiency.
Although the aforementioned methods are lightweight and efficient, the quality of SR reconstruction still requires significant improvement.

2.2. Attention Mechanism of Image SR

Researchers in the field of image super-resolution have adopted the attention mechanism, which was initially developed for natural language processing tasks [17,18], and has proven effective in image super-resolution.
Hu et al. [18] proposed using channel attention (CA), which assigns a weight to each feature channel based on its significance and improves the feature representation by amplifying the features with high weights and suppressing those with low weights. Hui et al. [15] enhanced the channel attention mechanism with contrast-aware channel attention (CCA), which assigns channel weights according to the sum of the standard deviation and the mean. Wang et al. [19] introduced efficient channel attention (ECA), which uses 1D convolution to efficiently capture dependencies across channels, to make the attention mechanism lighter.These attention mechanisms exhibit state-of-the-art performance in SR tasks [4,8,15].
Some studies have introduced spatial attention to enrich the feature map. Wang et al. [20] proposed an additional attention mechanism, non-local attention, which captures global context information by computing pixel-to-pixel dependencies. Nevertheless, this mechanism incurs a substantial computational overhead. To address this issue, Liu et al. [13] proposed enhanced spatial attention (ESA), which reduces the channel dimensions by employing a 1 × 1 convolutional layer followed by a stride convolution to expand the receptive field. The max pooling operation with a large window and stride then focuses on the feature’s crucial spatial information. EFDN [10] and BSRN [6] also demonstrate superior performance with ESA.
Guo et al. [21] proposed a novel linear attention mechanism named Large Kernel Attention (LKA) that utilizes the large receptive field of large convolutional kernels to achieve the effects of adaptability and long-range correlations similar to self-attention. The LKA attention mechanism has demonstrated excellent performance in various computer vision tasks [22,23]. However, the use of large convolutional kernels in LKA can introduce a significant computational burden. To address this, we decompose the large convolutional kernels in LKA into smaller ones, achieving results comparable to LKA while significantly reducing the computational requirements.

2.3. Factorized Convolution

Factorized convolution has emerged as a promising technique in efficient neural network design. It involves breaking down a standard convolution operation into multiple smaller convolution operations, typically aimed at reducing the computational complexity and model parameters. This technique has found widespread application in various computer vision tasks, including image classification, object detection, and semantic segmentation.
One common form of factorized convolution is depth-wise Separable Convolution, where a standard convolution layer is decomposed into two independent operations: depth-wise convolution and point-wise convolution. Depth-wise convolution independently filters each input channel spatially, while point-wise convolution combines the filtered outputs from each channel. This factorization significantly reduces the number of parameters, resulting in more efficient models.
Recent research has demonstrated the immense potential of factorized convolution in enhancing the efficiency of neural networks. For instance, MobileNet [24] introduced depth-wise Separable Convolution, creating lightweight models suitable for mobile devices. ERFNet [25] factorized 3 × 3 convolutions into 3 × 1 and 1 × 3 convolutions, achieving substantial performance improvements in semantic segmentation tasks. Subsequent studies like DABNet [26], LEDNet [27], and MSCFNet [28] have further improved upon this technique and successfully applied it to their respective tasks, further emphasizing the importance of factorized convolution in efficient network design.
While factorized convolution has been successful in tasks such as image classification and object detection, its application in enhancing the efficiency of super-resolution neural networks has been relatively limited. Despite the successes observed in image classification and object detection tasks, the untapped potential of factorized convolution in improving the efficiency of super-resolution neural networks remains largely unexplored.
To address this gap, we propose an innovative approach in this work, applying factorized convolution to super-resolution networks. Our method fully leverages the advantages of factorized convolution to create highly efficient and lightweight architectures capable of delivering high-quality image super-resolution results.

3. Method

3.1. Network Structure

Our CSINet intends to reconstruct HR images by leveraging the work of RFDN [13] and blueprint separable residual network (BSRN) [6]. Figure 2 depicts the architecture of CSINet, which comprises four modules: shallow feature extraction, multiple stacked feature aggregation residual group, dense feature fusion (DFF), and image reconstruction. The goal of the shallow feature extraction module is to extract low-level image features. The multiple stacked feature aggregation residual group is intended to aggregate and refine features from multiple scales. The dense feature fusion (DFF) module combines features from multiple scales, utilizing the attention mechanism to highlight important features and suppress irrelevant ones. The image reconstruction module then reconstructs the HR image based on the fused features.
Shallow Feature Extraction. Given a low-quality input image I LR R H × W × C , the shallow feature F 0 is extracted by a 3 × 3 convolutional layer. This process can be expressed as
F 0 = f c 3 × 3 ( I LR )
where f c n × m _ s k denotes an n × m convolutional layer with stride k for the shallow feature extraction; this convolution layer provides straightforward mapping from the input image space to a higher-dimensional feature space.
Multiple Stacked Feature Aggregation Residual Group. To extract deep features, we use a non-linear mapping module that consists of several stacked feature aggregation residual groups (FARGs). The output of the i-th FARG F i can be expressed as follows:
F i = FARG i ( F i 1 ) , i = 1 , 2 , , N
where FARG i ( · ) is the function of i-th FARG and the corresponding output denoted by F i . More details of FARG unit will be given in Section 3.3.
Dense Feature Fusion (DFF). To combine hierarchical features from all layers, the outputs of these FARGs are concatenated and sent to a DFF module consisting of a 1 × 1 convolution, a GELU, and a 3 × 3 convolution. The feature is then refined using an ESA attention module. This procedure can be described as
F DEF = Fusion ( Concat ( F 1 , F 2 , , F N ) ) F fused = ESA ( F DEF )
and
Fusion = f c 3 × 3 f GELU f c 1 × 1
where Concat ( F 1 , F 2 , , F N ) is the concatenation of features generated by all FARG units, F D F F is the aggregated feature, ESA denotes the spatial attention (ESA) (further details will be provided in Section 3.2), f GELU ( · ) is the Gaussian error linear unit activation function, and ∘ denotes function composition.
Image Reconstruction. The image reconstruction module consists of a 3 × 3 convolutional layer and a Pixel-Shuffle layer. The reconstruction stage is expressed as
F comb = F fused + F 0 I SR = f up , ps ( f c 3 × 3 ( F comb ) )
where I SR denotes the super-resolution result of the network, f up , ps indicates the Pixel Shuffle operation.
Loss Function. We utilize the L 1 loss to optimize the parameters of our CSINet model as
L = I HR I SR 1
where I SR is the super-resolution result of the network, and I HR denotes the corresponding high-resolution image.

3.2. Attention Modules

3.2.1. Efficient Large Kernel Attention (ELKA)

Guo et al. [21] introduced an innovative linear attention mechanism known as Large Kernel Attention (LKA), which leverages the expansive receptive field provided by large convolution kernels to attain adaptability and long-range correlation effects, akin to self-attention mechanisms. LKA has demonstrated remarkable efficacy, particularly in SR tasks [22,23]. Nonetheless, the utilization of large convolution kernels in LKA imposes a substantial computational burden.
To address this issue, we adopted two pivotal strategies. First, we decomposed the 2D convolution kernel in the deep convolution layer of LKA into a sequence of cascaded horizontal and vertical 1D convolution kernels. Specifically, a K × K spatial convolution was deconstructed into a K × 1 depth-wise convolution and a 1 × K depth-wise convolution. This decomposition effectively curtails the quadratic increase in the number of parameters in LKA as the convolution kernel size grows, all the while preserving performance quality.
Secondly, we introduced a 1 × 1 convolution layer both preceding and following the depth convolution operation, facilitating information interaction across channels. This groundbreaking module is denoted as ELKA. The overall architecture of ELKA is visually depicted in Figure 3.
The ELKA module consists of three parts: (1) Spatial local convolution: including two cascaded depth-wise convolution operations with convolution kernel sizes of 7 × 1 (DW-Conv7 × 1) and 1 × 7 (DW-Conv1 × 7). (2) Spatial global convolution: Contains two cascaded deep-wise dilated convolution operations with convolution kernel sizes of 9 × 1 (DW-D-Conv9 × 1) and 1 × 9 (DW-D-Conv1 × 9). (3) Two channel convolutions: These two channel convolutions are applied at the beginning and end of the module. The expression of ELKA operation is as follows:
F 1 ELKA = f GELU ( f c 1 × 1 ( F in ELKA ) ) F 2 ELKA = f dw 1 × 7 ( f dw 7 × 1 ( F 1 ELKA ) ) F 3 ELKA = f dw 1 × 9 _ r d ( f dw 9 × 1 _ r d ( F 2 ELKA ) ) F 4 ELKA = f c 1 × 1 ( F 3 ELKA ) F out ELKA = f c 1 × 1 ( F 1 ELKA F 4 ELKA )
where F out ELKA denotes the output of the ELKA module; f GELU ( · ) is the gelu activation function; f dw n × m indicates the n × m depth-wise convolution operation, f dw n × m _ r d is n × m depth-wise dilated convolution operation with dilated rate d, ⊙ indicates hadamard product.
Compared to the standard LKA design, ELKA can achieve comparable performance while exhibiting a lower computational complexity and memory.

3.2.2. Enhanced Spatial Attention

Enhanced Spatial Attention (ESA) is a lightweight and effective spatial attention mechanism [13], as shown in Figure 4. To reduce the computational cost, the ESA module first reduces the number of channels using 1 × 1 convolution. For enlarging the receptive field further, ESA first halves the size of the feature map using a 1 × 1 convolution with stride 2, and then adds a 7 × 7 max pooling layer with stride 3 to reduce the spatial dimension. Following a series of 3 × 3 convolutions to determine the interdependence of the feature map’s spatial dimensions, a bilinear interpolation is then used to restore the feature map to its original size and then concatenate the features obtained from the previous feature map. Then, a 1 × 1 convolutional layer is utilized to restore the number of channels of the feature map to its initial value. The attention mask is then generated by the sigmoid activation function and multiplied by the input features to produce an output feature map with long-distance dependence. Given the input characteristics F in ESA of ESA, the preceding operations can be described as follows:
F 1 ESA = f c 1 × 1 ( F in ESA ) F 2 ESA = Enlarge ( F 1 ESA ) F 3 ESA = F 1 ESA + f c 1 × 1 ( F 2 ESA ) w = f sigmoid ( f up , bi ( F 3 ESA ) ) F out ESA = F in ESA w
and
Enlarge = f upsample i f i , c 3 × 3 conv group f m 7 × 7 _ s 3 f c 1 × 1 _ s 2
where F out ESA denotes the output of the ESA module; f m n × m _ s k represents the n × m max pooling layer with stride k; f up , bi is the bilinear interpolation operation; f upsample is the upsampling operation; f sigmoid ( · ) is the sigmoid activation function; and ⊗ indicates the element-wise product.

3.3. Feature Aggregation Residual Group (FARG)

The feature distillation technique introduced in RFDN [13] has proven effective in reducing the number of parameters while improving performance. Nevertheless, recent studies [5] have indicated that eliminating the feature distillation branch can lead to a reduction in the runtime and computational cost. Motivated by the findings of [29], we have developed the Feature Aggregation Residual Group (FARG) architecture, which is depicted in Figure 2a.
FARG has been designed to be an efficient network module. It comprises two Cross-Scale Interaction Blocks (CSIB), a 3 × 3 depth-wise convolution layer, and employs the GELU activation function. To begin, FARG processes input features through a pair of CSIBs, a step critical for obtaining deep and robust feature representations. These channel feature enhancement blocks are instrumental in extracting and enriching information from the input features. Next, the channel features undergo convolution through 3 × 3 convolutional layers, further enhancing the feature representation. This step plays a crucial role in capturing spatial relationships and structural information between the features. Subsequently, the GELU activation function is applied for nonlinear transformation, introducing more complex nonlinear characteristics. This is valuable for enabling the model to better comprehend the intricacies of the data and extract abstract features. Finally, residual operations combine identity mapping with the output features, ensuring that the acquired features effectively integrate with the original input features to better preserve valuable information. This architectural design enhances the network module’s ability to learn and represent complex data features with greater effectiveness. The procedure of FARG can be expressed as
F 1 FARG = i CSIB i CSIB group ( F in FARG ) F out FARG = f c 3 × 3 ( f G E L U ( F 1 FARG ) + F in FARG )
where F in FARG and F out FARG are the input and output of the FARG, respectively; CSIB stands for the Cross-Scale Interaction Block, which will be introduced later; i CSIB i is a CSIB group as a sequence of CSIB blocks; here, two CSIBs are applied for our network; and f GELU ( · ) is the Gaussian error linear unit activation function.

3.4. Cross-Scale Interaction Block (CSIB)

To create an efficient architecture, we propose the efficient cross-scale interaction block (CSIB), inspired by the work of Romera [25], Wang [27], Li [26], and Gao [28]. The primary focus of the CSIB’s design is on cross-scale information interaction, taking into consideration the limitations of existing methods in terms of the feature representation capability and efficiency. The CSIB incorporates factorized depth-wise dilated convolutions and residual connections for efficient representation learning, as shown in Figure 5c. In contrast to the single-branch structure of the Non-bottleneck-1D module proposed by Romera [25] and the dual-branch structure of the SS-nbt module proposed by Wang [27], CSIB utilizes an effective cross-scale interaction technique to integrate cross-scale contextual information. This architecture is intended to strike a balance between accuracy and parameters, allowing for improved feature representation and enhanced computational efficiency.
Firstly, CSIB employs a 1 × 1 convolution layer to decrease the number of parameters and expedite the training process. Following [27,28], we employ a dual-branch structure to simultaneously extract local and multi-scale contextual information. Unlike SS-nbt [27], we replace the factorization convolution with depth-wise factorization convolution to further reduce the parameters in the first branch, which can extract local information. The second branch applies factorization convolution to the depth-wise dilated convolution in order to enlarge the receptive field, thereby capturing global context information. According to previous studies [26,28], dilated convolution may result in gridding artifacts; therefore, we employ depth-wise dilated convolutions with varying dilation rates in various CSIBs. To integrate the cross-scale contextual information of different branches, we perform an element-wise sum of the feature maps extracted by the 5 × 1 convolutions in the two branches and feed them to a subsequent 1 × 5 convolution in each branch. In this manner, the extracted feature maps from the two branches can interact.
Considering that concatenation operations are more effective than addition operations, we use concatenation to merge the convolution outputs of two branches. Due to the fact that the receptive fields of the two branches are different sizes, 1 × 1 convolution is used to promote the fusion of the contextual information extracted by the two branches, strengthen information interaction, and improve feature representation. Following this is an ELKA module for extracting distinguishing characteristics. Finally, the shortcut connection is utilized to preserve the previous functionality, and then we join this into the subsequent CSIB. The operations of CSIB can be expressed as
F 1 CSIB = f c 1 × 1 ( F in CSIB ) F 2 CSIB = f dw 5 × 1 ( F 1 CSIB ) + f dw 5 × 1 _ r d ( F 1 CSIB ) F 3 , L CSIB = f dw 1 × 5 ( F 2 CSIB ) F 3 , R CSIB = f dw 1 × 5 _ r d ( F 2 CSIB ) F 4 CSIB = ELKA ( f c 1 × 1 ( Concat ( F 3 , L CSIB , F 3 , R CSIB ) ) F out CSIB = F 4 CSIB + F in CSIB
where f dw n × m indicates the n × m depth-wise convolution operation, f dw n × m _ r d is n × m depth-wise dilated convolution operation with dilated rate d, ELKA indicates efficient large kernel attention module, Concat ( F 3 , L CSIB , F 3 , R CSIB ) is the concatenation of features generated by F 3 , L CSIB , F 3 , R CSIB .

4. Experiments

4.1. Experiment Setup

4.1.1. Datasets and Metrics

Following previous research [5,7,10,11,12,13], we train our models using the recently popularized dataset DIV2K [30] with 800 high-quality images. Five standard benchmark datasets are used to evaluate our models: Set5 [31], Set14 [32], BSD100 [33], Urban100 [34], and Manga109 [35]. To objectively assess the performance of our model, we convert the image to the YCbCr color space and compute the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) metrics on the luminance channel.
PSNR stands for Peak Signal-to-Noise Ratio and is a measure of image quality that compares the original image to the compressed or distorted image. It is defined as
PSNR = 10 log 10 M A X I 2 M S E
where M A X I is the maximum pixel value of the image, and M S E is the Mean Squared Error between the original and compressed/distorted images. Higher values of the PSNR indicate a better image quality.
SSIM stands for Structural Similarity Index and is a metric that compares the structural similarity of two images, taking into account the luminance, contrast, and structure. It is defined as:
SSIM ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
where x and y are the two images being compared, μ x and μ y are their respective means, σ x 2 and σ y 2 are their respective variances, and σ x y is their covariance. C 1 and C 2 are constants used to avoid instability when the means are close to zero. The SSIM value ranges between −1 and 1, where a value of 1 indicates perfect similarity.

4.1.2. Training Details

During the training phase, LR training images are generated by downsampling HR images with scaling factors ( × 2 , × 3 , and × 4 ) using bicubic interpolation in MATLAB R2017a. We apply random horizontal or vertical flips and 90° rotations to the training set. In each mini-batch, inputs consisting of 48 × 48 LR color patches are selected. The Adan optimizer will be used to train our model with the parameters β 1 = 0.98 , β 2 = 0.92 , β 3 = 0.99 , and with an initial learning rate of 1 × 10 3 . In the training stage, we use L1 to train our network for 1 × 10 6 iterations, and reduce the learning rate by half at 6 × 10 5 and 8 × 10 5 iterations. Subsequently, in the fine-tuning stage, we switch to the L2 to fine-tune our network with a learning rate of 2 × 10 5 , and a total of 1 × 10 5 iterations.
We replaced the 3 × 3 convolution in the FARG model with a 1 × 1 convolution, creating a smaller CSINet called CSINet-S. We trained CSINet-S using the DIV2K and Flickr2K datasets. During the training process, the input patch size is set to 64 × 64 and the mini-batch is set to 64. The Adan optimizer will be used to train our model with the parameters β 1 = 0.98 , β 2 = 0.92 , β 3 = 0.99 , and with an initial learning rate of 1 × 10 3 . We use L1 to train our network for 1 × 10 6 iterations, and reduce the learning rate by half at 6 × 10 5 and 8 × 10 5 iterations. Subsequently, in the fine-tuning stage, we switch to the L2 to fine-tune our network with a learning rate of 2 × 10 5 , and a total of 1 × 10 5 iterations.
The proposed networks are implemented using the PyTorch framework and trained on a single NVIDIA 3090 GPU.

4.2. Ablation Study

4.2.1. Effectiveness of Dilation Rate

In deep learning-based image super-resolution methods, the receptive field size of the network is an important factor that affects the ability of the network to capture spatial information from the input image. The dilation rate is a common way to adjust the receptive field size of a CNN. A larger dilation rate means a larger receptive field, which can capture more global contextual information, while a smaller dilation rate means a smaller receptive field, which can capture more local details.
As shown in Table 1, we conducted extensive experiments to investigate the effects of different dilation rates on the image super-resolution performance. Specifically, we adopted the concept from [26,28] and tested seven different dilation configurations. Our experimental results demonstrate that the choice of the dilation rate has a significant impact on the quality of the super-resolved images.
Among the tested dilation configurations, we found that setting the dilation rates to (1,3,3,5) consistently produces superior results across multiple benchmark datasets. These results are in line with previous studies that have also shown the effectiveness of large dilation rates in image super-resolution tasks.

4.2.2. Effectiveness of CSIB

The CSIB is intended to enhance the model’s reconstruction performance by effectively fusing multi-scale features from different branches. This is achieved through its parallel branching and cross-fusion structures. To evaluate the effectiveness of the CSIB, two similar modules were designed for comparative analysis (Figure 6).
The Multi-Branch Feature Fusion Block (MFFB), splits the input features into two branches using channel splitting and halving operations. The multi-scale contextual information is then extracted from these two branches using depth-wise factorized convolution. The Cascade Dilated Fusion Block (CDFB), employs a cascade structure of three 3 × 3 depth-wise dilated convolutions instead of the two-branch structure used in MFFB. Both of these modules were integrated into corresponding SR networks, named a Multi-Branch Feature Fusion Network (MFFNet) and Cascaded Dilated Fusion Network (CDFNet), respectively. Extensive experiments were conducted to evaluate the performance of these three SR networks, with the results shown in Table 2.
In terms of the reconstruction accuracy, the outcomes of these experiments clearly indicate that the CSIB is superior to both MFFB and CDFB. The CSIB achieved greater PSNR and SSIM values, indicating that the reconstructed images were more accurate. In addition, it did so with fewer parameters and at a lower computational cost, proving the efficacy of the interactive fusion structure in the SR reconstruction procedure. Compared to CDFB, CSIB not only requires fewer parameters, but also demonstrates a significant performance advantage in terms of reconstruction. This demonstrates the CSIB’s ability to not only have a large receptive field, but also effectively combine complementary information from multiple scales to improve the model’s representational capabilities.
The visual analysis of CDFB, MFFB, and CSIB is presented in Figure 7. As depicted in Figure 7a, CDFB shows promising results in recovering a portion of the butterfly’s streak profile, albeit with some blurring. In contrast, Figure 7b presents MFFB which stands out due to its ability to extract more details of the stripes. This enhanced performance is attributed to its effective use of multi-scale feature extraction modules, which facilitates the recovery of intricate details with remarkable precision.
Furthermore, the proposed CSIB, shown in Figure 7c, also utilizes multi-scale feature extraction modules, leading to a superior restoration performance when compared to the aforementioned models. CSIB excels in reconstructing high-frequency details and edge information with exceptional clarity, as evidenced in the results. The findings highlight the proficiency of CSIB in structural texture restoration and demonstrate the immense potential of deep learning models in image processing applications.

4.2.3. Effectiveness of Factorized Convolution

To validate the effectiveness of factorized convolution, we replaced it with regular convolution in CSIB, denoted as “w/RC”.
Upon reviewing the results in Table 3, it is clear that the inclusion of factorized convolution leads to a reduction of 14 K parameters and a decrease of 0.8 G FLOPs compared to regular convolution. Simultaneously, PSNR and SSIM exhibit improvements across all benchmark datasets. Furthermore, the inference time decreased by 1.24 ms. These findings indicate that the introduction of factorized convolution not only enhances the model’s lightweight characteristics but also contributes to significant performance improvements.

4.2.4. Effectiveness of ELKA and ESA

The ablation studies on the two attention modules—ELKA, and ESA—are presented in Table 4. The results indicate that ELKA is a highly effective module. We observed a significant decrease in the network performance when ELKA was removed, with a decrease of approximately 0.2 dB in the Set5 and Set14 datasets, and a decrease of over 0.4 dB in the Urban100 and Manga109 datasets. Furthermore, ESA has a positive impact on the model’s performance, as evidenced by a substantial decmidrule in the performance when ESA is removed.
These findings demonstrate that combining ELKA and ESA can effectively increase the model’s capacity. It is noteworthy that ELKA provides a more computationally efficient way to incorporate global information, while ESA modules can enhance the local feature representation. Thus, the combination of these attention modules offers a well-balanced and effective solution to improve the model’s performance.
To further observe the benefits produced by our ELKA module, we visualize the feature maps before and after ELKA for different FARGs, as shown in Figure 8. It can be observed that the ELKA module enhances high-frequency information, making the edges and structural textures in the output features clearer.

4.3. Comparison with the SOTA SR Methods

To verify the effectiveness of the proposed model, we compare our CSINet model with 14 lightweight state-of-the-art SISR methods, including SRCNN [1], VDSR [2], CARN [14], IDN [15], MAFFSRN [36], SMMR [7], IMDN [11], PAN [8], LAPAR-A [12], RFDN [13], Cross-SRN [37], FDIWN [38], RLFN [5], and BSRN [6]. The results of the comparisons are presented in Table 5. To assess the model’s size, we used two metrics: the number of parameters and the number of operations (Multi-Adds), calculated on a high-resolution image of 1280 × 720 . Our method achieved outstanding results on all the datasets with various scaling factors, outperforming most of the other state-of-the-art networks in both the PSNR and SSIM measurements. Despite having fewer parameters and Multi-Adds, our CSINet outperformed techniques such as LAPAR-A, RFDN, Cross-SRN, and even the RLFN, which was awarded second place in the sub-track2 (Overall Performance Track) of the NTIRE 2022 efficient super-resolution challenge. These results illustrate the effective balance between image quality and computational efficiency that our method achieves.
We have incorporated the Non-Reference Image Quality Evaluator (NIQE) into our evaluation metrics to provide a more comprehensive analysis of the performance of our model compared to other lightweight models, including VDSR, CARN, IMDN, PAN, EFDN, and RLFN, as shown in Table 6. In the comparison, we computed the NIQE scores for the outputs of our model and the aforementioned lightweight models. The NIQE score measures the naturalness of an image, with lower scores indicating a better image quality. Our model achieved comparable or slightly lower NIQE scores compared to these models, indicating that our model produces images with similar or slightly better naturalness. These results suggest that our model not only performs competitively in terms of traditional evaluation metrics such as PSNR and SSIM but also maintains or enhances the perceptual quality of the super-resolved images according to the NIQE score. It demonstrates the effectiveness of our lightweight model in preserving image quality while reducing the computational complexity.
The visual comparisons of our method with several state-of-the-art methods are presented in Figure 9, Figure 10 and Figure 11. The results demonstrate the superiority of our method in terms of the image quality.
For Set14, we compared the models’ ability to reconstruct the “baboon” and “monarch” images. Our findings suggest that while the SRCNN [1] and VDSR [2] models recovered most of the stripe contours, their reconstructions still exhibited blurriness. In contrast, our proposed model, CSINet, was able to reconstruct high-frequency details with greater clarity. For the “monarch” image, CSINet was also superior in reproducing the butterfly antennae with greater clarity.
On the BSD100 dataset, we evaluated the performance of the models on the “108005” and “148026” images. Our results indicate that Bicubic failed to reproduce the basic texture features when reconstructing the details of the stripes on the tiger. While other models, such as CARN [14], IMDN [11], PAN [8], and EFDN [10], could recover more stripe details, their reconstructed images still exhibited some blurriness. In contrast, CSINet was able to reconstruct high-frequency details with greater clarity, outperforming all the other models. For the “148026” image, CSINet also produced reconstructed images with clear texture and rich details, which were closer to the real images than the other models.
Finally, on the Urban100 dataset, we evaluated the models’ ability to restore the “img_092” image. Our results suggest that most of the models, except for Bicubic, could restore the horizontal stripes of the building facade but still exhibited some blurriness. In contrast, the reconstructed images from CSINet had clear texture and rich details, approaching perfection. Similarly, for the “img_062” image in the Urban100 test set, the reconstructed images using Bicubic, SRCNN [1], and VDSR [2] were severely distorted and blurry. While the reconstructed results using CARN [14], IMDN [11], PAN [8], EFDN [10], and E-RFDN [13] were slightly clearer, the glass window grids were distorted and deformed. In contrast, the reconstructed images using CSINet proposed in this study had clear texture and rich details, which were closer to the real images.
Overall, our subjective visual effect comparisons demonstrate that CSINet outperforms other state-of-the-art super-resolution models, providing high-frequency details that are clearer and closer to the real images.

4.4. Complexity Analysis

The runtime of a network is a crucial metric, even for lightweight SR algorithms. We conducted comparative experiments on the Set5 dataset ( × 4 ) to assess the reconstruction speeds of mainstream networks. The experiments were run on an NVIDIA 3090 GPU with 24 GB RAM. The test images had a spatial resolution of 64 × 64 pixels. After 10 repeated runs, the average inference times were obtained and are presented in Figure 12. It can be observed that our CSINet not only achieves the fastest reconstruction speed but also delivers the best reconstruction quality, demonstrating the significant advantages of our lightweight CSINet.
To further validate the lightweight nature of CSINet, we deployed it on the NVIDIA Jetson Xavier NX Developer Kit, known as one of the world’s smallest AI supercomputers for embedded MEC systems. We conducted experiments on real-world photos to evaluate the effectiveness of CSINet in the embedded MEC system. In these scenarios, ground-truth images and downsampling kernels were unavailable. As depicted in Figure 13, our method successfully reconstructs sharper and more accurate images compared to state-of-the-art approaches. This indicates that our lightweight model excels in achieving exceptional super-resolution performance, making it highly suitable for deployment in embedded MEC systems.

4.5. Discussions

The effectiveness of the proposed Cross-Scale Interaction Block (CSIB) is a key highlight of our study. CSIB stands out as a crucial component in enhancing the overall performance of CSINet.
Firstly, CSIB is meticulously designed for super-resolution (SISR), integrating cross-scale contextual information using depth-wise convolution and dilated convolution. This design choice proves effective in capturing and leveraging contextual details across different scales, contributing to improved image reconstruction.
Secondly, the incorporation of Efficient Lightweight Kernel Aggregation (ELKA) within CSIB further enhances the model’s representational capacity. ELKA plays a pivotal role in aggregating relevant features efficiently, contributing to the model’s ability to capture intricate details and patterns.
The experimental results underscore the effectiveness of CSIB. In comparison to scenarios where regular convolution is used, the inclusion of factorized convolution within CSIB leads to significant reductions in parameters and FLOPs while simultaneously improving PSNR, SSIM, and reducing the inference time. This indicates that CSIB not only reduces the model complexity but also positively impacts the image quality and computational efficiency.
In visual comparisons with state-of-the-art methods, CSINet equipped with CSIB excels in reconstructing high-frequency details with exceptional clarity. This suggests that the designed cross-scale interaction mechanism within CSIB plays a pivotal role in capturing and utilizing contextual information effectively, resulting in superior image reconstruction.
CSIB emerges as a crucial element contributing to the effectiveness of CSINet. Its innovative design and integration within the network significantly improve image quality, demonstrating the efficacy of the proposed cross-scale interaction strategy in the context of lightweight super-resolution.

5. Conclusions

In this paper, we introduce the Cross-Scale Interaction Network (CSINet), a novel architecture designed for lightweight image super-resolution (SISR). Specifically, we present a lightweight Cross-Scale Interaction Block (CSIB) tailored for SISR. This block is carefully crafted to integrate cross-scale contextual information using depth-wise convolution and dilated convolution, leading to an effective reduction in the model complexity. Additionally, the integration of Efficient Lightweight Kernel Aggregation (ELKA) enhances the model’s representational capacity. The proposed network is characterized by its lightweight nature, with only 366K parameters. Extensive experiments conducted on benchmark datasets validate that CSINet outperforms the majority of state-of-the-art lightweight SR methods. Remarkably, it achieves superior results with fewer parameters and Multi-Adds, underscoring its efficiency and effectiveness.
In future work, to enhance the applicability of CSINet in real-time scenarios, optimizing model parameters and the inference time will become crucial for achieving a more lightweight model. This optimization will remain a central focus in our ongoing research, with the goal of ensuring the seamless integration of CSINet into real-time application environments.

Author Contributions

Conceptualization, G.K.; methodology, G.K., S.-L.L. and Y.-F.L.; resources, Y.-F.L., H.Z. and Z.-Q.C.; software, G.K. and Z.-Q.C.; formal analysis, G.K., H.Z. and S.-L.L.; validation, J.-K.W.; writing—original draft, G.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Macau Science and Technology Development Funds [Grant number 0061/2020/A2]; this research was also funded by the Science and Technology of Social Development Program [Grant number 20211800904512,20231800935472], and Dongguan Sci-tech Commissoner Program [Grant number 20231800500352].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The available online experimental datasets in this paper are https://paperswithcode.com./ (accessed on 1 April 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 391–407. [Google Scholar]
  2. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  3. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  4. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  5. Kong, F.; Li, M.; Liu, S.; Liu, D.; He, J.; Bai, Y.; Chen, F.; Fu, L. Residual Local Feature Network for Efficient Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 766–776. [Google Scholar]
  6. Li, Z.; Liu, Y.; Chen, X.; Cai, H.; Gu, J.; Qiao, Y.; Dong, C. Blueprint Separable Residual Network for Efficient Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 833–843. [Google Scholar]
  7. Wang, L.; Dong, X.; Wang, Y.; Ying, X.; Lin, Z.; An, W.; Guo, Y. Exploring sparsity in image super-resolution for efficient inference. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4917–4926. [Google Scholar]
  8. Zhao, H.; Kong, X.; He, J.; Qiao, Y.; Dong, C. Efficient image super-resolution using pixel attention. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 56–72. [Google Scholar]
  9. Du, Z.; Liu, D.; Liu, J.; Tang, J.; Wu, G.; Fu, L. Fast and Memory-Efficient Network Towards Efficient Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 853–862. [Google Scholar]
  10. Wang, Y. Edge-Enhanced Feature Distillation Network for Efficient Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, New Orleans, LA, USA, 18–24 June 2022; pp. 777–785. [Google Scholar]
  11. Hui, Z.; Gao, X.; Yang, Y.; Wang, X. Lightweight Image Super-Resolution with Information Multi-distillation Network. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
  12. Li, W.; Zhou, K.; Qi, L.; Jiang, N.; Lu, J.; Jia, J. Lapar: Linearly-assembled pixel-adaptive regression network for single image super-resolution and beyond. Adv. Neural Inf. Process. Syst. 2020, 33, 20343–20355. [Google Scholar]
  13. Liu, J.; Tang, J.; Wu, G. Residual feature distillation network for lightweight image super-resolution. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 41–55. [Google Scholar]
  14. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
  15. Hui, Z.; Wang, X.; Gao, X. Fast and accurate single image super-resolution via information distillation network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 723–731. [Google Scholar]
  16. Chen, S.; Huang, K.; Li, B.; Xiong, D.; Jiang, H.; Claesen, L. Adaptive hybrid composition based super-resolution network via fine-grained channel pruning. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 119–135. [Google Scholar]
  17. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 3. [Google Scholar]
  18. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  19. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 11531–11539. [Google Scholar]
  20. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-Local Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  21. Guo, M.H.; Lu, C.Z.; Liu, Z.N.; Cheng, M.M.; Hu, S.M. Visual attention network. Comput. Vis. Media 2023, 9, 733–752. [Google Scholar] [CrossRef]
  22. Feng, H.; Wang, L.; Li, Y.; Du, A. LKASR: Large kernel attention for lightweight image super-resolution. Knowl.-Based Syst. 2022, 252, 109376. [Google Scholar] [CrossRef]
  23. Xie, C.; Zhang, X.; Li, L.; Meng, H.; Zhang, T.; Li, T.; Zhao, X. Large Kernel Distillation Network for Efficient Single Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; p. 1. [Google Scholar]
  24. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  25. Romera, E.; Alvarez, J.M.; Bergasa, L.M.; Arroyo, R. Erfnet: Efficient residual factorized convnet for real-time semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2017, 19, 263–272. [Google Scholar] [CrossRef]
  26. Li, G.; Yun, I.; Kim, J.; Kim, J. Dabnet: Depth-wise asymmetric bottleneck for real-time semantic segmentation. arXiv 2019, arXiv:1907.11357. [Google Scholar]
  27. Wang, Y.; Zhou, Q.; Liu, J.; Xiong, J.; Gao, G.; Wu, X.; Latecki, L.J. Lednet: A lightweight encoder-decoder network for real-time semantic segmentation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1860–1864. [Google Scholar]
  28. Gao, G.; Xu, G.; Yu, Y.; Xie, J.; Yang, J.; Yue, D. MSCFNet: A lightweight network with multi-scale context fusion for real-time semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2021, 23, 25489–25499. [Google Scholar] [CrossRef]
  29. Li, H.; Li, J.; Wei, H.; Liu, Z.; Zhan, Z.; Ren, Q. Slim-neck by GSConv: A better design paradigm of detector architectures for autonomous vehicles. arXiv 2022, arXiv:2206.02424. [Google Scholar]
  30. Agustsson, E.; Timofte, R. Ntire 2017 challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 126–135. [Google Scholar]
  31. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-Complexity Single-Image Super-Resolution Based on Nonnegative Neighbor Embedding British Machine Vision Conference. 2012. Available online: https://api.semanticscholar.org/CorpusID:5250573 (accessed on 6 February 2024).
  32. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the International Conference on Curves and Surfaces, Avignon, France, 24–30 June 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 711–730. [Google Scholar]
  33. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision. ICCV 2001, Vancouver, BC, Canada, 7–14 July 2001; IEEE: Piscataway, NJ, USA, 2001; Volume 2, pp. 416–423. [Google Scholar]
  34. Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  35. Matsui, Y.; Ito, K.; Aramaki, Y.; Fujimoto, A.; Ogawa, T.; Yamasaki, T.; Aizawa, K. Sketch-based manga retrieval using manga109 dataset. Multimed. Tools Appl. 2017, 76, 21811–21838. [Google Scholar] [CrossRef]
  36. Muqeet, A.; Hwang, J.; Yang, S.; Kang, J.; Kim, Y.; Bae, S.H. Multi-attention based ultra lightweight image super-resolution. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 103–118. [Google Scholar]
  37. Liu, Y.; Jia, Q.; Fan, X.; Wang, S.; Ma, S.; Gao, W. Cross-srn: Structure-preserving super-resolution network with cross convolution. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4927–4939. [Google Scholar] [CrossRef]
  38. Gao, G.; Li, W.; Li, J.; Wu, F.; Lu, H.; Yu, Y. Feature distillation interaction weighting network for lightweight image super-resolution. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 661–669. [Google Scholar]
Figure 1. Trade-off between performance and model complexity of other state-of-the-art lightweight models on the BSD100 dataset for × 4 SR. The CSINet achieves higher PSNR with fewer parameters.
Figure 1. Trade-off between performance and model complexity of other state-of-the-art lightweight models on the BSD100 dataset for × 4 SR. The CSINet achieves higher PSNR with fewer parameters.
Sensors 24 01135 g001
Figure 2. An overview of our CSINet network. (a) The architecture of CSINet network, (b) the details of the feature aggregation residual group (FARG).
Figure 2. An overview of our CSINet network. (a) The architecture of CSINet network, (b) the details of the feature aggregation residual group (FARG).
Sensors 24 01135 g002
Figure 3. Efficient Large Kernel Attention (ELKA).
Figure 3. Efficient Large Kernel Attention (ELKA).
Sensors 24 01135 g003
Figure 4. Enhanced Spatial Attention (ESA).
Figure 4. Enhanced Spatial Attention (ESA).
Sensors 24 01135 g004
Figure 5. The structures of different residual modules. (a) Non-bt-1D [25]. (b) SS-nbt [27]. (c) Our CSIB. ‘C’ is the number of the input channels, ‘DW’ indicates a depth-wise convolution, and ‘DW-D’ denotes a depth-wise dilated convolution.
Figure 5. The structures of different residual modules. (a) Non-bt-1D [25]. (b) SS-nbt [27]. (c) Our CSIB. ‘C’ is the number of the input channels, ‘DW’ indicates a depth-wise convolution, and ‘DW-D’ denotes a depth-wise dilated convolution.
Sensors 24 01135 g005
Figure 6. Two different CSIB modules. (a) MFFB. (b) CDFB. ‘C’ is the number of the input channels, ‘DW’ indicates a depth-wise convolution, and ‘DW-D’ denotes a depth-wise dilated convolution.
Figure 6. Two different CSIB modules. (a) MFFB. (b) CDFB. ‘C’ is the number of the input channels, ‘DW’ indicates a depth-wise convolution, and ‘DW-D’ denotes a depth-wise dilated convolution.
Sensors 24 01135 g006
Figure 7. Visualized feature maps processed by different convolution designs. (a) Input feature. (b) Feature processed by the CDFB. (c) Feature processed by the MFFB. (d) Output feature of CSIB.
Figure 7. Visualized feature maps processed by different convolution designs. (a) Input feature. (b) Feature processed by the CDFB. (c) Feature processed by the MFFB. (d) Output feature of CSIB.
Sensors 24 01135 g007
Figure 8. Visualized feature maps of the four FARGs. (a) Visualization feature maps of the four FARGs before ELKA. (b) Visualization feature maps of the four FARGs after ELKA. The values are calculated by averaging the feature maps and normalized in range 0 to 1.
Figure 8. Visualized feature maps of the four FARGs. (a) Visualization feature maps of the four FARGs before ELKA. (b) Visualization feature maps of the four FARGs after ELKA. The values are calculated by averaging the feature maps and normalized in range 0 to 1.
Sensors 24 01135 g008
Figure 9. Visual comparison of the Set 14 dataset for × 4 SR.
Figure 9. Visual comparison of the Set 14 dataset for × 4 SR.
Sensors 24 01135 g009
Figure 10. Visual comparison on the BSD100 dataset for × 4 SR.
Figure 10. Visual comparison on the BSD100 dataset for × 4 SR.
Sensors 24 01135 g010
Figure 11. Visual comparison on the Urban100 dataset for × 4 SR.
Figure 11. Visual comparison on the Urban100 dataset for × 4 SR.
Sensors 24 01135 g011
Figure 12. Average running time on Set5 dataset for × 4 SR.
Figure 12. Average running time on Set5 dataset for × 4 SR.
Sensors 24 01135 g012
Figure 13. Comparison of super-resolution results on real-world photos; CSINet outperforms state-of-the-art methods with embedded MEC system.
Figure 13. Comparison of super-resolution results on real-world photos; CSINet outperforms state-of-the-art methods with embedded MEC system.
Sensors 24 01135 g013
Table 1. Investigation of different dilate rates. ‘R’ denotes the dilation rates of each depth-wise convolution. These results were recorded after 1 × 10 6 iterations without pre-training and fine-tuning. The best are color red.
Table 1. Investigation of different dilate rates. ‘R’ denotes the dilation rates of each depth-wise convolution. These results were recorded after 1 × 10 6 iterations without pre-training and fine-tuning. The best are color red.
Dilation RateSet5Set14BSD100Urban100Manga109
PNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIM
R = (1,2,2,4)32.26/0.896328.68/0.784527.64/0.740426.22/0.791630.58/0.9102
R = (1,2,2,6)32.26/0.896128.66/0.784027.65/0.740326.23/0.791530.59/0.9100
R = (1,2,4,6)32.29/0.896128.69/0.784327.64/0.740426.22/0.791630.55/0.9099
R = (1,3,5,7)32.32/0.896528.67/0.784127.63/0.740326.21/0.791630.56/0.9099
R = (1,3,5,5)32.29/0.896328.69/0.784427.64/0.740226.18/0.790030.58/0.9101
R = (1,3,3,5)32.34/0.896528.68/0.784527.64/0.740526.23/0.791830.58/0.9103
Table 2. Quantitative comparison of three distinct approaches to × 4 SR: MFFNet, CDFNet, and the proposed CSINet. These results were recorded after 1 × 10 6 iterations without pre-training and fine-tuning. The best results are color red.
Table 2. Quantitative comparison of three distinct approaches to × 4 SR: MFFNet, CDFNet, and the proposed CSINet. These results were recorded after 1 × 10 6 iterations without pre-training and fine-tuning. The best results are color red.
MethodParamsMulti-AddsSet5Set14BSD100Urban100Manga109
PNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIM
MFFNet311 K17.3 G32.17/0.894828.55/0.781227.53/0.736626.01/0.783430.32/0.9068
CDFNet343 K20.5 G32.15/0.894428.62/0.782727.60/0.738926.02/0.785930.32/0.9075
CSINet366 K20.5 G32.34/0.896528.68/0.784527.64/0.740526.23/0.791830.58/0.9103
Table 3. The results pertaining to the inclusion of factorized convolution in CSIB are presented. “w/RC” denotes the scenario where factorized convolution is replaced with regular convolution, while “w/FC” signifies our model using factorized convolution. The inference time is calculated on Set5 with a scaling factor of × 4 . The experiments were executed using an NVIDIA 3090 GPU. The best results are color red.
Table 3. The results pertaining to the inclusion of factorized convolution in CSIB are presented. “w/RC” denotes the scenario where factorized convolution is replaced with regular convolution, while “w/FC” signifies our model using factorized convolution. The inference time is calculated on Set5 with a scaling factor of × 4 . The experiments were executed using an NVIDIA 3090 GPU. The best results are color red.
MethodParamsMulti-AddsAve. TimeSet5Set14BSD100Urban100Manga109
PNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIM
w/RC380 K21.3 G10.08 ms32.27/0.896128.66/0.783727.63/0.739826.15/0.788630.60/0.9098
w/FC366 K20.5 G8.84 ms32.34/0.896528.68/0.784527.64/0.740526.23/0.791830.58/0.9103
Table 4. Comparison of the number of parameters, Multi-Adds, and mean values of PSNR obtained without ELKA and without ESA and our CSINet on five datasets for × 4 SR. These results were recorded after 1 × 10 6 iterations without pre-training and fine-tuning.The best results are color red.
Table 4. Comparison of the number of parameters, Multi-Adds, and mean values of PSNR obtained without ELKA and without ESA and our CSINet on five datasets for × 4 SR. These results were recorded after 1 × 10 6 iterations without pre-training and fine-tuning.The best results are color red.
MethodParamsMulti-AddsSet5Set14BSD100Urban100Manga109
PNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIM
w/o ELKA273 K15.3 G32.01/0.892728.47/0.779627.51/0.776825.81/0.776830.05/0.9030
w/o ESA343 K20.4 G32.26/0.896028.65/0.784027.63/0.740126.23/0.791430.55/0.9101
CSINet366 K20.5 G32.34/0.896528.68/0.784527.64/0.740526.23/0.791830.58/0.9103
Table 5. Quantitative comparisons of state-of-the art SR algorithm on five datasets. The best and the second best results are color red and blue, respectively. “Multi-Adds” are computed with a 720p HR image.
Table 5. Quantitative comparisons of state-of-the art SR algorithm on five datasets. The best and the second best results are color red and blue, respectively. “Multi-Adds” are computed with a 720p HR image.
MethodsScaleParamsMulti-AddsSet5Set14BSD100Urban100Manga109
PNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIMPNSR/SSIM
Bicubic × 2 --33.66/0.929930.24/0.868829.56/0.843126.88/0.840330.80/0.9339
SRCNN [1] × 2 8 K52.7 G36.66/0.954232.42/0.906331.36/0.887929.50/0.894635.60/0.9663
VDSR [2] × 2 666 K612.6 G37.53/0.958733.03/0.912431.90/0.896030.76/0.914037.22/0.9750
CARN [14] × 2 1592 K222.8 G37.76/0.959033.52/0.916632.09/0.897831.92/0.925638.36/0.9765
IDN [15] × 2 553 K124.6 G37.83/0.960033.30/0.914832.08/0.898531.27/0.919638.01/0.9749
MAFSSRN [36] × 2 402 K77.2 G37.97/0.960333.49/0.917032.14/0.899431.96/0.9268-
SMMR [7] × 2 985 K131.6 G38.00/0.960133.64/0.917932.17/0.899032.19/0.928438.76/0.9771
IMDN [11] × 2 694 K158.8 G38.00/0.960533.63/0.917732.19/0.899632.17/0.928338.88/0.9774
PAN [8] × 2 261 K70.5 G38.00/0.960533.59/0.918132.18/0.899732.01/0.927338.70/0.9773
LAPAR-A [12] × 2 548 K171.0 G38.01/0.960533.62/0.918332.19/0.899932.10/0.928338.67/0.9772
RFDN [13] × 2 534 K95 G38.05/0.960633.68/0.918432.16/0.899432.12/0.927838.88/0.9773
Cross-SRN [37] × 2 --38.03/0.960633.62/0.918032.19/0.899732.28/0.929038.75/0.92773
FDIWN-M [38] × 2 -------
RFLN [5] × 2 527 K-38.07/0.960733.72/0.918732.22/0.900032.33/0.9299-
BSRN [6] × 2 332 K73.0 G38.10/0.961033.74/0.919332.24/0.900632.34/0.930339.14/0.9782
CSINet-S (ours) × 2 248 K54.6 G38.06/0.960833.82/0.920032.26/0.900932.40/0.931339.08/0.9780
CSINet (ours) × 2 348 K77.7 G38.08/0.960833.77/0.920532.27/0.900932.45/0.931839.00/0.9779
Bicubic × 3 --30.39/0.868227.55/0.774227.21/0.738524.46/0.734926.95/0.8556
SRCNN [1] × 3 8 K52.7 G32.75/0.909029.30/0.821528.41/0.786326.24/0.798930.48/0.9117
VDSR [2] × 3 666 K612.6 G33.66/0.921329.77/0.831428.82/0.797627.14/0.827932.01/0.9340
CARN [14] × 3 1592 K118.8 G34.29/0.925530.29/0.840729.06/0.803428.06/0.849333.50/0.9440
IDN [15] × 3 553 K57.0 G34.11/0.925329.99/0.835428.95/0.801327.42/0.835932.71/0.9381
MAFSSRN [36] × 3 418 K34.2 G34.32/0.926930.35/0.842929.09/0.805228.13/0.8521-
SMMR [7] × 3 993 K67.8 G34.40/0.927030.33/0.841229.10/0.805028.25/0.853633.68/0.9445
IMDN [11] × 3 703 K71.5 G34.36/0.927030.32/0.841729.09/0.804628.17/0.851933.61/0.9445
PAN [8] × 3 261 K39 G34.40/0.927130.36/0.842329.11/0.805028.11/0.851133.61/0.9448
LAPAR-A [12] × 3 544 K114 G34.36/0.926730.34/0.842129.11/0.805428.15/0.852333.51/0.9441
RFDN [13] × 3 541 K42.2 G34.41/0.927330.34/0.842029.09/0.805028.21/0.852533.67/0.9449
Cross-SRN [37] × 3 --32.43/0.927530.33/0.841729.09/0.805028.23/0.853533.65/0.9448
FDIWN-M [38] × 3 446 K35.9 G34.46/0.927430.35/0.842329.10/0.805128.16/0.8528-
RFLN [5] × 3 -------
BSRN [6] × 3 340 K33.3 G32.46/0.927730.47/0.844929.18/0.806828.39/0.856734.05/0.9471
CSINet-S (ours) × 3 255 K25.1 G34.47/0.927530.46/0.844929.18/0.807628.37/0.857333.91/0.9464
CSINet (ours) × 3 356 K35.3 G34.49/0.927930.49/0.845329.19/0.807728.40/0.857733.93/0.9464
Bicubic × 4 --28.42/0.810426.00/0.702725.96/0.667523.14/0.657724.89/0.7866
SRCNN [1] × 4 57 K52.7 G30.48/0.862627.50/0.751326.90/0.710124.52/0.722127.58/0.8555
VDSR [2] × 4 666 K612.6 G31.35/0.883828.01/0.767427.29/0.725125.18/0.752428.83/0.8770
CARN [14] × 4 1592 K90.9 G32.13/0.893728.60/0.780627.58/0.734926.07/0.783730.47/0.9084
IDN [15] × 4 553 K32.3 G31.82/0.890328.25/0.773027.41/0.729725.41/0.763229.41/0.8942
MAFSSRN [36] × 4 441 K19.3 G32.18/0.894828.58/0.781227.57/0.736126.04/0.7848-
SMMR [7] × 4 1006 K41.6 G32.12/0.893228.55/0.780827.55/0.735126.11/0.786830.54/0.9085
IMDN [11] × 4 715 K40.9 G32.21/0.894828.58/0.781127.56/0.735326.04/0.783830.45/0.9075
PAN [8] × 4 272 K28.2 G32.13/0.894828.61/0.782227.59/0.736326.11/0.785430.51/0.9095
LAPAR-A [12] × 4 548 K94 G32.15/0.894428.61/0.781827.61/0.736626.14/0.787130.42/0.9074
RFDN [13] × 4 550 K23.9 G32.24/0.895228.61/0.781927.57/0.736026.11/0.785830.58/0.9089
Cross-SRN [37] × 4 --32.24/0.895428.59/0.781727.58/0.736426.17/0.788130.53/0.9088
FDIWN-M [38] × 4 454 K19.6 G32.17/0.894128.55/0.780627.58/0.736426.02/0.7844-
RFLN [5] × 4 543 K-32.24/0.895228.62/0.781327.60/0.736426.17/0.7877-
BSRN [6] × 4 352 K19.4 G32.35/0.896228.73/0.784727.65/0.738726.27/0.790830.84/0.9123
CSINet-S (ours) × 4 266 K14.7 G32.24/0.895928.72/0.783927.64/0.738526.22/0.790130.68/0.9097
CSINet (ours) × 4 366 K20.5 G32.37/0.897128.78/0.785727.69/0.739826.35/0.793230.85/0.9117
Table 6. The average NIQE for × 4 SR. Red indicates the best performance. The best results are color red.
Table 6. The average NIQE for × 4 SR. Red indicates the best performance. The best results are color red.
MethodScaleSet5Set14BSD100Urban100
VDSR [2] × 4 8.54586.90626.98906.2648
CARN [14] × 4 7.14666.28806.57945.7105
IMDN [11] × 4 6.88196.29016.54135.6889
EFDN [10] × 4 7.0746.16426.54605.6847
RFLN [5] × 4 7.28056.20116.57555.7244
CSINet × 4 6.78456.16136.51935.7888
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ke, G.; Lo, S.-L.; Zou, H.; Liu, Y.-F.; Chen, Z.-Q.; Wang, J.-K. CSINet: A Cross-Scale Interaction Network for Lightweight Image Super-Resolution. Sensors 2024, 24, 1135. https://doi.org/10.3390/s24041135

AMA Style

Ke G, Lo S-L, Zou H, Liu Y-F, Chen Z-Q, Wang J-K. CSINet: A Cross-Scale Interaction Network for Lightweight Image Super-Resolution. Sensors. 2024; 24(4):1135. https://doi.org/10.3390/s24041135

Chicago/Turabian Style

Ke, Gang, Sio-Long Lo, Hua Zou, Yi-Feng Liu, Zhen-Qiang Chen, and Jing-Kai Wang. 2024. "CSINet: A Cross-Scale Interaction Network for Lightweight Image Super-Resolution" Sensors 24, no. 4: 1135. https://doi.org/10.3390/s24041135

APA Style

Ke, G., Lo, S. -L., Zou, H., Liu, Y. -F., Chen, Z. -Q., & Wang, J. -K. (2024). CSINet: A Cross-Scale Interaction Network for Lightweight Image Super-Resolution. Sensors, 24(4), 1135. https://doi.org/10.3390/s24041135

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop