Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
A TCN-Linear Hybrid Model for Chaotic Time Series Forecasting
Previous Article in Journal
MV–MR: Multi-Views and Multi-Representations for Self-Supervised Learning and Knowledge Distillation
Previous Article in Special Issue
Unveiling the Future of Human and Machine Coding: A Survey of End-to-End Learned Image Compression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Convolutional Neural Network-Based Quantization Method for Block Compressed Sensing of Images

1
School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, China
2
North Automatic Control Technology Institute, Taiyuan 030006, China
3
Beijing Institute of Astronautical Systems Engineering, Beijing 100076, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(6), 468; https://doi.org/10.3390/e26060468
Submission received: 4 March 2024 / Revised: 25 May 2024 / Accepted: 28 May 2024 / Published: 29 May 2024
(This article belongs to the Special Issue Information Theory and Coding for Image/Video Processing)

Abstract

:
Block compressed sensing (BCS) is a promising method for resource-constrained image/video coding applications. However, the quantization of BCS measurements has posed a challenge, leading to significant quantization errors and encoding redundancy. In this paper, we propose a quantization method for BCS measurements using convolutional neural networks (CNN). The quantization process maps measurements to quantized data that follow a uniform distribution based on the measurements’ distribution, which aims to maximize the amount of information carried by the quantized data. The dequantization process restores the quantized data to data that conform to the measurements’ distribution. The restored data are then modified by the correlation information of the measurements drawn from the quantized data, with the goal of minimizing the quantization errors. The proposed method uses CNNs to construct quantization and dequantization processes, and the networks are trained jointly. The distribution parameters of each block are used as side information, which is quantized with 1 bit by the same method. Extensive experiments on four public datasets showed that, compared with uniform quantization and entropy coding, the proposed method can improve the PSNR by an average of 0.48 dB without using entropy coding when the compression bit rate is 0.1 bpp.

1. Introduction

Compressed sensing (CS) [1,2,3,4] is very suitable for resource-constrained applications due to its low complexity [5,6,7,8]. In CS, it is necessary to quantize real-valued CS measurements. The CS framework with the quantization process is called quantized compressed sensing (QCS) [8].
The QCS primarily focuses on optimizing the encoder or decoder to enhance the quality of signal reconstruction for the commonly used quantization methods. The exploration of advanced compression techniques in the realm of block-based compressed sensing (BCS) has been a focal point in contemporary research. One such advancement is the application of Differential Pulse Code Modulation (DPCM) for quantized block-based compressed sensing of images. This strategy leverages DPCM’s efficiency in exploiting spatial correlations within image blocks to minimize quantization errors, demonstrating improved bit-rate efficiency without sacrificing reconstruction quality [9]. However, the quantized measurements must undergo entropy coding to attain ideal performance. Ref. [10] proposes a progressive quantization method, which essentially improves the encoding and decoding strategies while still utilizing uniform scalar quantization. The reconstruction algorithm [11,12,13,14] has been a primary focal point in CS for optimizing the decoder.
In order to improve the quantization performance, vector quantization has been used to quantify the CS measurements [15,16]. Subsequently, to further amplify the efficiency of the vector quantization technique, ref. [17] has leveraged deep neural networks. Due to the high computational complexity of vector quantization, scalar quantization is more suitable for CS measurements. In data compression theory, the scalar quantizer that performs entropy coding on quantized output data is usually called an entropy-constrained quantizer or entropy-coded quantizer [18]. When quantization error is used as a distortion measurement, the uniform scalar quantizer is the optimal entropy coding quantizer in rate-distortion performance [19,20]. In other words, for the BCS measurements, the rate-distortion performance of the uniform or non-uniform scalar quantization methods will be inferior to the joint performance of uniform quantization and entropy coding. Therefore, in current research on CS for images, the CS measurements are usually quantified using the uniform scalar quantization method [21], and the quantized measurements are encoded by entropy coding to improve the compression performance [22,23]. However, since the computational cost involved in entropy coding is usually high [23,24], using entropy coding will reduce the low-complexity advantage of the CS encoder.
There are two main ways to improve rate-distortion performance. The first way is to reduce the bitrate while keeping the distortion constant, while the second way is to reduce the distortion while keeping the bitrate constant. Using entropy coding on quantized data is the first way. Without considering entropy coding, the second way is the only method to improve rate-distortion performance. Moreover, there are strategies aimed at enhancing the compression efficiency of CS. For example, ref. [25] introduces a novel application of Zadoff–Chu sequences, renowned for their excellent autocorrelation properties. By utilizing this sequence in the measurement matrix, sparsity in the compressive domain is enhanced, leading to improved recovery accuracy for quantized CS data. Ref. [26] explores the use of Discrete Fourier Transform (DFT) for measurement, enabling parallel processing capabilities and enhancing the efficiency of block compressive sensing. This strategy capitalizes on the computational advantages of FFT to accelerate the CS process while maintaining reconstruction fidelity. While many approaches can improve compression efficiency at the encoding stage [25,26,27,28], they are not necessarily applicable or effective for QCS.
While the field of block compressed sensing (BCS) has witnessed significant advancements in recent years, traditional quantization techniques employed in image/video coding applications continue to face pivotal challenges. Specifically, uniform quantization, a common choice due to its simplicity and compatibility with entropy coding, often incurs substantial quantization errors, compromising the fidelity of the reconstructed images. Moreover, it tends to overlook the inherent structure and correlations within the data, leading to encoding redundancy and inefficiency. On the other hand, Lloyd–Max quantization, although theoretically optimal for minimizing mean squared error, necessitates computationally intensive offline training and struggles to adapt dynamically to varying image characteristics. Furthermore, the reliance on entropy coding as a supplementary step to mitigate the loss from quantization adds to the computational burden and complexity of the encoding process.
In light of these challenges, our work introduces a novel convolutional neural network (CNN)-based quantization method specifically designed to overcome the drawbacks of traditional approaches. By leveraging the power of deep learning, our method transcends the uniformity imposed by classic quantizers, achieving a more nuanced mapping of measurements that closely follows their underlying distribution. This adaptive quantization, coupled with a sophisticated dequantization mechanism that harnesses correlation from quantized data, significantly reduces quantization errors without resorting to additional entropy coding.
In this paper, we propose a quantization method of BCS measurements that can reduce distortion while maintaining the bitrate. At the end of encoding, the proposed method models the measurements’ distribution of each image block. Subsequently, it quantizes the measurements of data that conform to a uniform distribution based on the distribution model. The proposed method uses the distribution parameters of each image block as side information of the encoder and adopts the same strategy to quantize the side information with 1 bit. At the end of decoding, the proposed method first restores the quantized data to data that conform to the distribution of the original measurements, then extracts the correlation information of the measurements from the quantized data of adjacent blocks to correct the measurements. Before the dequantization of the measurements, the same strategy is used to dequantize the side information. All quantization and dequantization processes are implemented as convolutional neural networks (CNN), and all networks are jointly learned offline. The CS coding structure based on the proposed method is shown in Figure 1.
The main contributions of this article are as follows:
  • A quantization method of BCS measurement based on CNN is proposed. The proposed method constructs and jointly trains the CNN of the quantization and dequantization processes, which aims to maximize the coding output entropy and minimize the quantization error.
  • A quantization process based on measurements’ distribution is proposed. Based on the properties of the cumulative distribution function (CDF), a neural network model was constructed to map measurements to the quantized data following a uniform distribution, which maximizes the amount of information carried in the quantized data. An activation function with a constrained output range was designed to reduce the computational complexity of the network’s activation function.
  • A dequantization process based on the neighborhood information of measurements is proposed. The inverse process of quantization is used as a module to restore the quantized data to data that conform to the distribution of the original measurements. Furthermore, an information correction module is introduced to extract correlation information from multiple quantized values for correcting the measurements. The two modules are used to improve the quality of the dequantized measurements through residual connection.
  • The distribution parameters of the block measurements are used as side information, which is quantized with 1 bit by the same quantization process.
While conventional approaches such as uniform quantization and Lloyd–Max quantization with entropy coding have been widely employed, they often introduce significant quantization errors and encoding inefficiencies. Our work diverges from these methodologies by introducing a CNN-based quantization strategy that not only maps measurements to a uniformly distributed quantized space to maximize the information content but also incorporates a novel dequantization process that leverages correlations from the quantized data to minimize reconstruction errors. This innovative method bypasses the need for entropy coding, offering a more efficient and adaptive solution for BCS applications.
In comparison to uniform quantization, which assigns equal intervals to the entire dynamic range, and Lloyd–Max quantization, known for minimizing the mean squared error but requiring complex optimization, our CNN-based approach dynamically adapts to the underlying data distribution. Unlike entropy coding methods, which reduce redundancy at the expense of computational complexity, our method directly optimizes the quantization process through learning, achieving superior performance without additional encoding steps.
The core of our method lies in the design of the CNN architecture, which jointly learns the quantization and dequantization processes. This contrasts with traditional quantization techniques that rely on predetermined, static decision boundaries. Our network specifically tailors the quantization levels based on the input data’s statistical properties, ensuring a closer match to the original signal characteristics. Additionally, the utilization of correlation information from the quantized data for post-processing further distinguishes our method, leading to reduced quantization artifacts.
The rest of this paper is organized as follows. Section 2 introduces the proposed method, which mainly includes the BCS quantization process, parameter estimation, parameter quantization and dequantization, and the BCS dequantization process. Section 3 presents the experimental results. The conclusion is given in Section 4.

2. Proposed Method

In this paper, we aim to improve the amount of information of quantized data and the information extraction ability of the dequantization process. The proposed method mainly includes the following aspects: (1) BCS quantization process, (2) parameter estimation, (3) parameter quantization and dequantization, and (4) BCS dequantization process.

2.1. BCS Quantization Process

Ref. [29] demonstrates that a random variable’s CDF can transform its data distribution into a uniform distribution. The random variables are usually modeled by probability density functions (PDF) to describe their distributions. It is difficult to obtain a closed-form expression for the CDF due to the need for integral calculations involving the PDF. Due to the strong function approximation ability of feedforward neural networks, we propose utilizing a feedforward neural network to model the CDF of the CS measurements.
Assuming that P ( y ) and F ( y ) are the PDF and CDF of the measurement variable y , respectively, P ( y ) and F ( y ) need to satisfy the following conditions:
F ( y ) 0 F ( y ) 1 P ( y ) = F ( y ) x 0
Compared with a three-layer feedforward neural network, a four-layer feedforward neural network can accurately establish the relationship between input and target variables with fewer hidden neurons [30]. When employing a four-layer feedforward neural network to build the CDF of the measurements, it can be represented as:
u 1 = y u 2 = g 1 W 1 u 1 + d 1 u 3 = g 2 W 2 u 2 + d 2 F ( y ) = g 3 W 3 u 3 + d 3
where g 1 , g 2 , g 3 represent the activation functions, and W 1 L 1 × 1 , W 2 L 2 × L 1 , W 3 1 × L 2 , d 1 L 1 × 1 , d 2 L 2 × 1 , d 3 1 × 1 denote the model parameters.
Based on Equation (1), it is necessary to ensure that the output values of F ( y ) fall within [0, 1]. In model (2), the activation function g 3 determines the output range of F ( y ) . Among the commonly used activation functions, the sigmoid function can ensure that the output values fall within [0, 1]. However, the sigmoid is a highly complex nonlinear activation function. We propose a rectified linear activation function with a limited output range, which can be expressed as:
G ( x ) = 1 x α x 2 α + 1 2 α x α 0 x α
where α is a finite constant greater than 0. The curve graph and gradient curve of G ( x ) are shown in Figure 2.
To improve the adaptiveness of the activation function, we take α as a learnable parameter of the CDF model. Because activation functions usually do not come with trainable parameters, we transform Equation (3) as:
g 3 ( z ) = 1 z 1 z 2 + 1 2 1 z 1 0 z 1
By comparing Equations (3) and (4), we obtain G ( x ) = g 3 x α . Let β = 1 a , and then the neural network model of CDF can be expressed as:
F ( y ) = g 3 β W 3 g 2 W 2 g 1 ( W 1 y + d 1 ) + d 2 + d 3
where β , and g 1 , g 2 are the LeakyReLU activation function [31].
In BCS, all measurements are usually stored in the form of a matrix, which can be expressed as:
Y = y 1 , , y N B = Φ x 1 , , x N B
where Y M × N B represents the measurements’ matrix, Φ represents the measurement matrix, M represents the number of measurements for each image block, N B represents the number of image blocks, x i represents the pixel value vector of the i-th block, and y i represents the measurements’ vector of the i-th block.
When the input becomes a matrix, we can convert the feedforward neural network to a CNN. The number of convolutional layers in the CNN equals the number of parameter layers in the feedforward neural network. The number of channels in the convolutional network is equal to the number of neurons in the hidden layer. Since the feedforward neural network model has a single input and output, the kernel size of the convolutional network is 1 × 1. Considering both quality and complexity, we set the output feature of the intermediate convolutional layer to six channels. The network structure diagram of CNN based on model (5) is shown in Figure 3.
In Figure 3, “Conv” denotes convolutions, the numbers above “Conv” indicate the kernel size, “LeakyReLU” and “ g 3 ” represent the activation functions used in the current convolutional layer, and the rectangular boxes represent the output feature maps of the convolutional layer. The numbers below the rectangular boxes indicate the number of channels. To balance the computational cost and fitting performance of the model (5), we adopt L 1 = L 2 = 6 in this paper.
Due to variations between image blocks, the distribution parameters and CDF of different block measurements also differ. However, the model (5) does not account for the influence of distribution parameters, which limits its adaptability.
When using a Gaussian measurement matrix, the measurements of the same image block approximately follow a Gaussian distribution. Assuming that the measurement variable y j of the j-th image block follows a Gaussian distribution N ( y j ; μ j , σ j ) , it can be transformed into a standard normal distribution N ( z ; 0 , 1 ) through Equation (7), which can be expressed as:
z = y j μ j σ j
where μ j and σ j represent the location parameter (mean) and scale parameter (standard deviation), respectively. For the measurements’ matrix Y , Equation (7) can be expressed as:
Z = Y U . / Λ
where U = μ j represents the matrix of position parameters, Λ = σ j represents the matrix of scale parameters, and . / represents the element-wise division of matrices.
After removing the impact of distribution parameters through Equation (8), the same CNN can transform the measurements of all blocks into a uniform distribution.
In order to reduce the computational burden of parameter extraction, the matrix Y 0 M 0 × N B of a small number of measurements is used for extracting the distribution parameters. The proposed BCS quantization process is shown in Figure 4. In Figure 4, “Repmat” represents repeat copies of an array, which is used to copy the parameters into the same size as the measurements’ matrix. Because the measurements of an image block have the same position and scale parameters, the size of the position and scale parameters are 1 × N B .

2.2. Parameter Estimation

In general, the expressions of the mean (position parameter) μ and variance (scale parameter) σ 2 are as follows:
μ j = i y j , i N σ j 2 = 1 N i y j , i μ 2
Since using partial measurements to estimate distribution parameters may introduce some errors, we use a neural network to estimate the distribution parameters. According to the definition of convolution, Equation (9) can be expressed as:
μ j = i y j , i N = [ y j , 1 , y j , 2 , ] W u σ j 2 = 1 N i y j , i μ 2 = [ [ y j , 1 , y j , 2 , ] [ y j , 1 , y j , 2 , ] W μ [ y j , 1 , y j , 2 , ] [ y j , 1 , y j , 2 , ] W μ ] W μ
where denotes the convolution operation, . / represents the element-wise multiplication of matrices, and W μ = 1 N , , 1 N denotes the convolution kernel used for computing the mean.
Equation (10) shows that the mean and variance can be estimated using convolutions. Based on Equation (10), we construct CNNs to estimate the position and scale parameters, as shown in Figure 5 and Figure 6, respectively. Since the mean and standard deviation functions are relatively simple in form, three channels are used for the middle convolution layer.

2.3. BCS Dequantization Process

The information of the quantized data is not only present in a single quantized value but also exists among multiple quantized values. Typically, the dequantization process at the decoding end is the inverse of the quantization process at the encoding end. However, the proposed quantization process and its inverse process only operate on individual quantized values. If dequantization only utilizes the inverse process of the proposed quantization process, it cannot use the correlated information among the multiple quantized values. Therefore, we propose adding a measurements’ information correction module to extract the measurement correction from the multiple neighboring quantized values. The dequantization process is shown in Figure 7.
The quantization process is built based on the CDF and its inverse process can be constructed with the inverse function of the CDF. We adopt the same network architecture for the inverse process module. Since the similarity between adjacent image blocks is significant, we use convolution kernels of 1 × 3 to extract the compensation information from the neighborhood measurements. In order to improve the information extraction, five convolution layers are used in the measurement compensation information module and the inverse process of the proposed quantization. The network structure of the inverse process of the proposed quantization is shown in Figure 8. The network structure of the measurement compensation information module is illustrated in Figure 9.

2.4. Parameter Quantization and Dequantization

Due to the high similarity between adjacent image blocks, the distribution parameters exhibit some similarity. Therefore, we use the same quantization and dequantization processes to quantize and dequantize the distribution parameters. To reduce the extra bits of the side information, we use 1 bit to quantize the side information. The network structures of quantization and dequantization for the parameters are shown in Figure 10 and Figure 11, respectively.

2.5. Local Normalization and Loss Function

2.5.1. Local Normalization of Measurements’ Matrix

The element in the i-th row and j-th column of the measurements’ matrix can be expressed as:
Y i , j = Φ i x j = k = 1 N Φ i , k x k , j
where Φ i represents the i-th row of the measurement matrix Φ . The grayscale values of the images are typically represented by 256 levels, the pixel values satisfy the following:
0 x k , j 255 , x k , j
Combining Equations (11) and (12), we have:
255 k = 1 N min Φ i , k , 0 k = 1 N Φ i , k x k , j 255 k = 1 N max Φ i , k , 0
Based on Equation (13), we take 255 k = 1 N min Φ i , k , 0 and 255 k = 1 N max Φ i , k , 0 as the minimum and maximum values of Y i , j , which can be expressed as:
Y i , j m a x = 255 k = 1 N max Φ i , k , 0 Y i , j m i n = 255 k = 1 N min Φ i , k , 0
Based on Equation (14), each measurement can be normalized, which can be expressed as:
Y ˜ i , j = Y i , j Y i , j m i n Y i , j m a x Y i , j m i n
According to Equation (14), it is known that the same row of Y shares the same maximum and minimum values. Therefore, Equation (15) is referred to as the local normalization method for the measurements’ matrix.
Since Equation (14) only requires computation based on the measurement matrix, the row normalization method does not require transmitting the maximum and minimum value. Equation (15) transforms all measurements into real numbers between 0 and 1, and the input and output of the CNNs in the quantization and the dequantization processes are also real numbers between 0 and 1. Therefore, the quantization of the measurements’ matrix can be expressed as:
Q Y = Round F q n e t Y ˜ 2 b 1
where b represents the bit-depth of quantization, and F q n e t represents the CNN in the quantization process.
At the decoding end, the output value of the dequantization network needs to be denormalized, which can be represented as:
Y ^ = F d q n e t Q 1 Q F q n e t Y ˜ Y i , j m a x Y i , j m i n + Y i , j m i n
where F d q n e t represents the CNN in the dequantization process.

2.5.2. Loss Function

The main objective of the proposed method is to optimize the amount of information retained in the quantized data while minimizing the quantization error. The average information of data is usually measured by information entropy. However, since the information entropy depends on probability statistics, it cannot propagate the gradient. In this paper, we propose a continuous function to estimate the information entropy when training the network.
The entropy of the quantized measurements’ matrix Y can be represented as follows:
H ( Y ) = k P ( s k ) log 2 P ( s k )
where P ( s k ) represents the probability of the symbol s k in Y , which requires counting the number of occurrences of symbol s k .
We can count the numbers greater than s k in the measurements’ matrix by using a step function, which is defined as follows:
h x = 1 x 0 0 x < 0
The numbers greater than s k in the measurements’ matrix can be represented as:
i j h Y i , j s k
The step function h x is not differentiable at 0. To ensure differentiability, we use the sigmoid activation function to approximate h x . The approximate function can be represented as follows:
h x s i g m o i d ( η x )
When η + , we can get:
s i g m o i d η Y i , j s k + ε = 1 Y i , j s k 0 Y i , j < s k
where 0 < ε < 1 is used to ensure that the function maps to 1 when Y i , j = s k , and we set ε = 0.5. The parameter η can be used to adjust the variation of the sigmoid function. Based on practical experience, setting η = 64 can make the output values of the sigmoid function approach 0 or 1 as much as possible.
Based on Equation (22), the probability greater than s k in the measurements’ matrix can be represented as:
P ( Y s k ) i M j N B s i g m o i d η Y i , j s k + 0.5 M N B
Assuming that s k + 1 > s k , we have:
P ( s k ) P ( Y s k ) P ( Y s k + 1 ) i M j N B s i g m o i d η Y i , j s k + 0.5 M N B i M j N B s i g m o i d η Y i , j s k + 1 + 0.5 M N B
Based on the above, the computation of information entropy can be estimated as:
H ( Y ) = k P ( Y s k ) P ( Y s k + 1 ) log 2 P ( Y s k ) P ( Y s k + 1 )
To simultaneously minimize the quantization error and maximize the information entropy, the objective function is composed of two parts. The first part is the mean square error (MSE) between the CS measurements before and after quantization, and the second part is the information entropy of the quantized measurements. The loss function is as follows:
L o s s = M S E ( Y , Y ^ ) λ H ( Y )
where λ > 0 is a parameter that controls the importance of the information entropy of quantized data. Y ^ represents the dequantized measurements.

3. Results

In this section, we present various experimental results that validate the performance of our method. The proposed method is primarily implemented through CNNs, which requires the collection of a training dataset for network training. The training dataset comprises 200 training images taken from the BSDS500 dataset [32]. Each image has been cropped into grayscale images of size 256 × 256 with a stride of 60 pixels. The block size utilized in BCS is set at 16 × 16. The samples and labels of the training data set are both a matrix of BCS measurements for each image. The matrix used to collect the measurements has a sampling rate of 0.8, so the trained network can be applied to any measurements’ matrix with a sampling rate lower than 0.8. Each block typically requires at least ten measurements to reconstruct an image from the BCS measurements efficiently, so we take M 0 = 10 as the number of the partial measurements.
All CNNs were implemented using the Pytorch framework. We trained the CNNs of the quantization and dequantization processes together. The batch size was set to 32, with the optimization process performed using the Adam algorithm, initialized with a learning rate of 0.001. After the initial training of 10,000 epochs, the learning rate was reduced by a factor of 10, and all networks were trained for an additional 20,000 epochs. The training process was conducted on a server powered by an Intel Xeon CPU, a Nvidia RTX 2080Ti GPU with 11 GB of memory, and 128 GB of DDR4 RAM. The test images consisted of an APC, aerial, airplane, airport, building, moon surface, tank, and truck, as illustrated in Figure 12. Publicly available datasets such as Set5 [33] (5 images), Set11 [34] (11 images), Set14 [35] (14 images), and BSD68 [36] (68 images) were also employed. All experiments were conducted on an Intel Core i5-8300H CPU @ 2.30GHz, and the proposed method’s performance was measured using the PSNR of the reconstructed images.

3.1. Analysis of Measurement Reconstruction Results

In this section, we analyzed the number of reconstruction levels and the quantization errors of the quantization method.
The current studies on improving CS quantization methods typically focus on sparse signals, but these methods are not suitable for images with a large number of elements. To ensure a low complexity of the encoder, the BCS encoder usually uses uniform quantization and entropy coding to process the BCS measurements. In addition, uniform quantizer is considered as the optimal quantizer for entropy-coded quantization in data compression theory [23,24], which is why BCS encoders tend to use uniform quantization and entropy coding. Currently, the most advanced quantization techniques for BCS of images are believed to be the prediction quantization method [9] and the progressive quantization method [10]. However, they essentially improve the coding strategy, while the quantization method employed is still uniform quantization, which can explore improvements using the commonly used quantization methods and the proposed method. To simplify the experimental process, this paper only compares the quantization techniques, using entropy-coded uniform quantization, µ-law quantization [37,38], and Lloyd–max quantization [39,40] as benchmarks. The entropy-coded uniform quantization method refers to the use of entropy coding after performing uniform quantization on the measurements. The codebook for Lloyd–Max quantization was obtained through offline training.
In scalar quantization methods, the number of reconstruction levels is typically equal to the number of quantization levels, as shown in Table 1. Table 2 shows the number of reconstructed levels of the proposed method for eight test images at a measurement rate of 0.2. Comparison between Table 1 and Table 2 shows that the proposed method have more different elements in the dequantized measurements. This is mainly because the proposed method utilizes the information of multiple quantized values for the dequantization, which gives the proposed method the advantage of many-to-many mapping. Moreover, each row of the measurements’ matrix adopts different maximum and minimum values for local denormalization. The local normalization approach also increases the number of reconstruction levels in accordance with the increase in measurement rate.
The greater the number of reconstruction levels of dequantized measurements, the greater the quantization error can be reduced. Table 3, Table 4 and Table 5 display the MSE of the various quantization methods for the measurements quantized with 3-bit, 6-bit, and 8-bit, respectively.
Table 3 shows that when using 3-bit quantization, the proposed method reduces the MSE by 788.07, 670.48, and 585.35 compared with uniform quantization, µ-law quantization, and Lloyd–Max quantization, respectively. Similarly, in Table 4, the proposed method accomplishes a reduced MSE by 8.25, 6.77, and 10.66 when 6-bit quantization is employed. Table 5 reveals that when 8-bit quantization is applied, the proposed method reduces the MSE by 0.4665, 0.3765, and 1.8162 compared with uniform quantization, µ-law quantization, and Lloyd–Max quantization, respectively. Table 3, Table 4 and Table 5 demonstrate that the proposed method has a significantly lesser MSE than other quantization methods.

3.2. Analysis of the Impact of Entropy Loss Constraints

In this section, we analyzed the effect of entropy constraint. The parameter λ in the loss function determines the degree of entropy constraint. In order to analyze the appropriate value of λ , we only select a few common values for training the CNNs of the quantization and dequantization processes. Table 6 shows the MSE of the dequantized measurements and the information entropy of the quantized measurements when 8-bit is used to quantize the measurements of the BSDS500 dataset.
It can be seen from Table 6 that the entropy constraint can increase the information entropy of the quantized measurements but it has a slight impact on reducing the MSE of the dequantized measurements. When λ = 0.05 , the MSE of the dequantized measurements is the smallest. Therefore, when training the proposed network, we use λ = 0.05 .
In addition, it can be observed that the entropies of the quantified measurements are very close to the bit-depth. Some images may not be compressed when using a fixed code table for entropy coding. In other words, the measurements quantized by the proposed method do not need entropy coding.

3.3. Analysis of the Impact of the Measurement Information Correction Module

In this section, we analyzed the impact of the information correction module on the dequantization process. Table 7 shows the quantization performance of the BSDS500 dataset when different information correction modules are used in the proposed method. In Table 7, Contrast Scheme 1 did not use a measurement information correction module. Contrast Scheme 2 used a measurement information correction module composed of three convolutional layers. Contrast Scheme 3 used a measurement information correction module composed of six convolutional layers.
Table 7 reveals that the proposed method has significant advantages over the µ-law quantization method. Compared with Contrast Scheme 1, the MSE of Contrast Scheme 2 is reduced by 0.0271, while its entropy is increased by 0.067. Similarly, the MSE of Contrast Scheme 3 is reduced by 0.0457 and the entropy is increased by 0.1616. These results illustrate that the information correction module effectively improves the quantization performance. Furthermore, the information correction module exhibits a stronger correction capability with the increase of convolutional layers.

3.4. Rate-Distortion Performance Comparison

In this section, we compared the rate-distortion performance of the proposed method with the entropy-coded uniform quantization, µ-law quantization, and Lloyd–Max quantization. The entropy-coded uniform quantization method refers to the use of entropy coding after performing uniform quantization on the measurements, which is expressed by “uniform quantization + entropy coding” in this paper. When drawing the rate-distortion curve, we traverse multiple quantization bit-depths and sampling rates to encode and decode the test images. Then, we choose the optimum Bitrate-PSNR points and connect them with a line. The bit-depth adopts seven values in {2, 3, 4, …, 8}, and the sampling rate chooses 77 values in {0.04, 0.05, 0.06, …, 0.8}. The image reconstruction algorithm used is the BCS-SPL-DCT algorithm [41]. When calculating the bitrate of “uniform quantization + entropy coding,” the average codeword length of entropy coding is replaced by information entropy. Figure 13 shows the PSNR curve of the eight test images.
In Figure 13, the proposed method has the best rate-distortion performance on all eight test images, particularly for the aerial, building, and tank images. The PSNR curve of “uniform quantization + entropy coding” is better than the µ-law and Lloyd–Max quantization methods. This observation confirms that the existing quantizers without entropy coding have inferior rate-distortion performance compared with “uniform quantization + entropy coding.”
Figure 14 shows the reconstructed images of the eight test images with different methods at a compression bit rate of 0.1. The Lloyd–Max quantization approach generates an adaptive quantization dictionary for each image. We do not count the bits of the quantization dictionary in the compression bit rate of the Lloyd–Max quantization method. Therefore, the results of the Lloyd–Max quantization method in Figure 14 are equivalent to the optimal results of the conventional quantizer.
As shown in Figure 14, the proposed method exhibits the best visual effect and PSNR, followed by “uniform quantization + entropy coding” and Lloyd–Max quantization. Compared with “uniform quantization + entropy coding,” for the eight test images, the PSNR of the proposed method increased by 0.65 dB, 0.44 dB, 1.97 dB, 0.02 dB, 0.46 dB, 0.09 dB, 0.37 dB, and 0.29 dB, respectively. Compared with the Lloyd–Max quantization, for the eight test images, the PSNR of the proposed method increased by 2.1 dB, 0.75 dB, 1.8 dB, 0.28 dB, 0.78 dB, 1.55 dB, 1.76 dB, and 1.53 dB, respectively.
The four quantization methods were also tested on the four test image datasets, and the average PSNR curves are shown in Figure 15.
In Figure 15, the average PSNR is the mean of the PSNR of the reconstructed images at a given bit rate for all images in the dataset. The PSNRs at a given bit rate are obtained by linear interpolation from the Bitrate-PSNR curve for each image. The given bit rates are set to {0.1, 0.2, …, 1 bpp}. For datasets Set5, Set11, Set14, and BSD68, the average PSNR curve of the proposed method is better than “uniform quantization + entropy coding,” µ-law quantization and Lloyd–Max quantization. Particularly, at a low bit rate (around 0.1 bpp), the proposed method’s PSNR is much higher than the other methods.
The datasets SunHays80 [42] and Urban100 [43] have been extended for testing (all data are converted to grayscale images with 256 × 256). The quality of reconstruction is evaluated by the peak signal to noise ratio (PSNR) and the structural similarity (SSIM) between the reconstructed image and the original image. Table 8 shows the PSNRs and SSIMs of the four datasets at a bit rate of 0.1 bpp. Table 9 shows the PSNRs and SSIMs of the four datasets at a bit rate of 0.2 bpp.
For all images of the six datasets, when the bitrate is set to 0.1 bpp, the proposed method, “uniform quantization + entropy coding,” µ-law quantization, and Lloyd–Max quantization achieve average PSNRs of 19.69 dB, 19.24 dB, 17.54 dB, and 18.67 dB, respectively. Compared with “uniform quantization + entropy coding,” the proposed method improves the PSNR by an average of 0.45 dB without entropy coding. The proposed method, “uniform quantization + entropy coding,” µ-law quantization, and Lloyd–Max quantization achieve average SSIMs of 0.1855, 0.1739, 0.1408, and 0.1547 respectively. Compared with “uniform quantization + entropy coding,” the proposed method improves the SSIM by an average of 0.0116 without entropy coding.
For all images of the six datasets, when the bitrate is set to 0.2 bpp, the proposed method, “uniform quantization + entropy coding,” µ-law quantization, and Lloyd–Max quantization achieve average PSNRs of 21.27 dB, 21.09 dB, 20.88 dB, and 20.93 dB, respectively. Compared with “uniform quantization + entropy coding,” the proposed method improves the PSNR by an average of 0.18 dB without entropy coding. The proposed method, “uniform quantization + entropy coding,” µ-law quantization, and Lloyd–Max quantization achieve average SSIMs of 0.2738, 0.2683, 0.2543, and 0.2567, respectively. Compared with “uniform quantization + entropy coding,” the proposed method improves the SSIM by an average of 0.0055 without entropy coding.
Across all images in the six datasets, the proposed method demonstrates superior performance compared to standard uniform quantization combined with “uniform quantization + entropy coding,” as well as the μ-law and Lloyd–Max quantization schemes.

3.5. Analysis of Computational Complexity

On the encoding side, the calculation of the proposed quantization method involves four networks: the position parameter estimation network, the scale parameter estimation network, the parameter quantization network, and the measurement quantization network. The network structure of the position parameter estimation network and scale parameter estimation network are identical, as shown in Table 10. Similarly, the network structures of the parameter quantization and measurement quantization are identical, as shown in Table 11.
The position and scale parameters are derived from partial measurements Y 0 10 × N B . According to Table 10, convolutional layer 1 requires around 10 × N B × 3 multiplications and 10 × N B × 3 additions, convolutional layer 2 requires 3 × N B × 3 multiplications and 3 × N B × 3 additions, and convolutional layer 3 requires 3 × N B multiplications and 3 × N B additions. In total, location parameter estimation and scale parameter estimation need about 84 N B times multiplications, 84 N B times additions, and 12 N B times LeakyReLU operations.
According to Table 11, to quantize a parameter or measurement, the quantization network typically requires 48 times multiplications, 48 times additions, 12 times LeakyReLU operations, and one time g 3 operation. For the measurements’ matrix Y M × N B , the numbers of measurements and parameters that need to be quantized are M × N B and 2 N B , respectively. In total, it is necessary to compute ( 2 + M ) 48 N B times multiplications, ( 2 + M ) 48 N B times additions, ( 2 + M ) N B times LeakyReLU, and ( 2 + M ) N B times g 3 .
The proposed method requires approximately the same number of multiplication and addition operations, and the activation function to be computed only involves linear operations. Since addition is much faster than multiplication in practical operations, we only compared the number of multiplication operations. With an image size of 256 × 256 and a block size of 16 × 16, the total number N B of blocks would be 256. Assuming that the measurement rate of BCS is 0.1, each block obtains 26 measurements. Each measurement needs about 256 times multiplications and 255 times additions, and the calculation of measurements is about 26 × 256 N B times multiplications and 26 × 255 N B times additions. The proposed quantization method requires 660 N B times multiplications, 660 N B times additions, 156 N B times LeakyReLUs, and 12 N B times g 3 . Compared with the calculation for BCS measurements, the calculation of the proposed quantization process is about 9.92% of that of the BCS measurements.

4. Conclusions

In this paper, we propose a BCS measurement quantization method based on CNNs. The method uses a CNN based on measurements’ CDF to map BCS measurements into quantized data following a uniform distribution. The block measurements’ distribution parameters are quantized as the side information of the encoder. In dequantization, a CNN based on information correction is designed by using the correlations between the BCS measurements. The proposed method uses partial measurements to extract parameter features and jointly optimizes the CNNs of the quantization and dequantization. The experimental results indicate that the proposed quantization method has better performance than joint uniform quantization and entropy coding.
In the future, investigating the performance of our method under adverse conditions, such as noisy or blurred inputs, is critical for real-world deployment. While the CNNs have shown remarkable performance, their computational and memory requirements are points for consideration. Exploring lightweight CNN architectures, pruning techniques, and hardware acceleration strategies would be instrumental in making our method more practical for resource-constrained devices.

Author Contributions

Conceptualization, J.G.; Methodology, J.G., Q.C. and Z.W.; Software, J.G. and Q.C.; Validation, J.G., Q.C., W.Z. and Z.W.; Formal analysis, Q.C., W.Z. and Z.W.; Resources, J.G.; Data curation, J.G., Q.C., W.Z. and Z.W.; Writing—original draft, Q.C.; Writing—review & editing, J.G., W.Z. and Z.W.; Visualization, Q.C. and Z.W.; Project administration, J.G.; Funding acquisition, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [State Key Laboratory of Explosion Science and Safety Protection] grant number [QNKT23-07] And the APC was funded by [Beijing Institute of Technology].

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Candès, E.J.; Romberg, J.; Tao, T. Robust Uncertainty Principles: Exact Signal Frequency Information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar] [CrossRef]
  2. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  3. Candès, E.J. Compressive sampling. Proc. Int. Congr. Math. Madrid 2006, 2006, 1433–1452. Available online: http://www.ems-ph.org/books/show_abstract.php?proj_nr=39&vol=3&rank=69%5Cnpapers3://publication/doi/10.4171/022-3/69 (accessed on 1 January 2024).
  4. Foucart, S.; Rauhut, H. A Mathematical Introduction to Compressive Sensing; Birkhäuser: Basel, Switzerland, 2013. [Google Scholar]
  5. Wakin, M.B.; Laska, J.N.; Duarte, M.F.; Baron, D.; Sarvotham, S.; Takhar, D.; Kelly, K.F.; Baraniuk, R.G. An architecture for compressive imaging. In Proceedings of the 2006 International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 1273–1276. [Google Scholar] [CrossRef]
  6. Romberg, J. Imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 14–20. [Google Scholar] [CrossRef]
  7. Li, X.; Lan, X.; Yang, M.; Xue, J.; Zheng, N. Efficient lossy compression for compressive sensing acquisition of images in compressive sensing imaging systems. Sensors 2014, 14, 23398–23418. [Google Scholar] [CrossRef] [PubMed]
  8. Goyal, V.K.; Fletcher, A.K.; Rangan, S. Compressive sampling and lossy compression. IEEE Signal Process. Mag. 2008, 25, 48–56. [Google Scholar] [CrossRef]
  9. Mun, S.; Fowler, J.E. DPCM for quantized block-based compressed sensing of images. In Proceedings of the 2012 Proceedings of the 20th European Signal Processing Conference (EUSIPCO), Bucharest, Romania, 27–31 August 2012; pp. 1424–1428. [Google Scholar]
  10. Wang, L.; Wu, X.; Shi, G. Binned progressive quantization for compressive sensing. IEEE Trans. Image Process. 2012, 21, 2980–2990. [Google Scholar] [CrossRef] [PubMed]
  11. Yang, Y.; Sun, J.; Li, H.; Xu, Z. ADMM-CSNet: A Deep Learning Approach for Image Compressive Sensing. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 521–538. [Google Scholar] [CrossRef] [PubMed]
  12. Zha, Z.; Yuan, X.; Wen, B.; Zhou, J.; Zhu, C. Group Sparsity Residual Constraint with Non-Local Priors for Image Restoration. IEEE Trans. Image Process. 2020, 29, 8960–8975. [Google Scholar] [CrossRef] [PubMed]
  13. Zhou, S.; He, Y.; Liu, Y.; Li, C.; Zhang, J. Multi-Channel Deep Networks for Block-Based Image Compressive Sensing. IEEE Trans. Multimed. 2021, 23, 2627–2640. [Google Scholar] [CrossRef]
  14. Cui, W.; Liu, S.; Jiang, F.; Zhao, D. Image Compressed Sensing Using Non-local Neural Network. IEEE Trans. Multimed. 2021, 25, 816–830. [Google Scholar] [CrossRef]
  15. Shirazinia, A.; Chatterjee, S.; Skoglund, M. Joint source-channel vector quantization for compressed sensing. IEEE Trans. Signal Process. 2014, 62, 3667–3681. [Google Scholar] [CrossRef]
  16. Leinonen, M.; Codreanu, M.; Juntti, M. Distributed Distortion-Rate Optimized Compressed Sensing in Wireless Sensor Networks. IEEE Trans. Commun. 2018, 66, 1609–1623. [Google Scholar] [CrossRef]
  17. Leinonen, M.; Codreanu, M. Low-Complexity Vector Quantized Compressed Sensing via Deep Neural Networks. IEEE Open J. Commun. Soc. 2020, 1, 1278–1294. [Google Scholar] [CrossRef]
  18. Sayood, K. Introduction to Data Compression; Elsevier: Amsterdam, The Netherlands, 2012. [Google Scholar] [CrossRef]
  19. Farvardin, N.; Modestino, J.W. Optimum Quantizer Performance for a Class of Non-Gaussian Memoryless Sources. IEEE Trans. Inf. Theory 1984, 30, 485–497. [Google Scholar] [CrossRef]
  20. Gish, H.; Pierce, J.N. Asymptotically Efficient Quantizing. IEEE Trans. Inf. Theory 1968, 14, 676–683. [Google Scholar] [CrossRef]
  21. Cui, W.; Jiang, F.; Gao, X.; Zhang, S.; Zhao, D. An efficient deep quantized compressed sensing coding framework of natural images. In Proceedings of the MM’18: Proceedings of the 26th ACM international conference on Multimedia, Seoul Republic of Korea, Seoul, Republic of Korea, 22–26 October 2018; pp. 1777–1785. [Google Scholar] [CrossRef]
  22. Chen, Z.; Hou, X.; Shao, L.; Gong, C.; Qian, X.; Huang, Y.; Wang, S. Compressive sensing multi-layer residual coefficients for image coding. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 1109–1120. [Google Scholar] [CrossRef]
  23. Unde, A.S.; Deepthi, P.P. Rate–distortion analysis of structured sensing matrices for block compressive sensing of images. Signal Process. Image Commun. 2018, 65, 115–127. [Google Scholar] [CrossRef]
  24. Goyal, V.K. Theoretical foundations of transform coding. IEEE Signal Process. Mag. 2001, 18, 9–21. [Google Scholar] [CrossRef]
  25. Xue, L.; Qiu, W.; Wang, Y.; Wang, Z. An Image-Based Quantized Compressive Sensing Scheme Using Zadoff–Chu Measurement Matrix. Sensors 2023, 23, 1016. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, Z.; Jiang, Y.; Chen, S. Image parallel block compressive sensing scheme using DFT measurement matrix. Multimed Tools Appl. 2023, 82, 21561–21583. [Google Scholar] [CrossRef]
  27. Kusetogullari, H.; Yavariabdi, A. Evolutionary multiobjective multiple description wavelet based image coding in the presence of mixed noise in images. Appl. Soft Comput. 2018, 73, 1039–1052. [Google Scholar] [CrossRef]
  28. Kusetogullari, H. Dynamic Multiple Description Wavelet Based Image Coding Using Enhanced Particle Swarm Optimization. IEEE J. Sel. Top. Signal Process. 2015, 9, 125–138. [Google Scholar] [CrossRef]
  29. Pharr, M.; Jakob, W.; Humphreys, G. Physically Based Rendering: From Theory to Implementation, 4th ed. Available online: https://pbr-book.org/4ed/contents (accessed on 1 January 2024).
  30. Tamura, S.; Tateishi, M. Capabilities of a four-layered feedforward neural network: Four layers versus three. IEEE Trans. Neural Networks 1997, 8, 251–255. [Google Scholar] [CrossRef] [PubMed]
  31. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; p. 30. [Google Scholar]
  32. Arbeláez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 898–916. [Google Scholar] [CrossRef] [PubMed]
  33. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Morel, M.L.A. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the 23rd British Machine Vision Conference, Surrey, UK, 3–7 September 2012. [Google Scholar] [CrossRef]
  34. Kulkarni, K.; Lohit, S.; Turaga, P.; Kerviche, R.; Ashok, A. ReconNet: Non-iterative reconstruction of images from compressively sensed measurements. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar] [CrossRef]
  35. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Lecture Notes in Computer Science (Including Subseries Lecture Notes Artificial Intelligence Lecture Notes Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2012; pp. 711–730. [Google Scholar] [CrossRef]
  36. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001, Vancouver, BC, Canada, 7–14 July 2001; pp. 416–423. [Google Scholar] [CrossRef]
  37. Gray, R.M.; Neuhoff, D.L. Quantization. IEEE Trans. Inf. Theory 1998, 44, 2325–2383. [Google Scholar] [CrossRef]
  38. Jacques, L.; Hammond, D.K.; Fadili, M.J. Stabilizing nonuniformly quantized compressed sensing with scalar companders. IEEE Trans. Inf. Theory 2013, 59, 7969–7984. [Google Scholar] [CrossRef]
  39. Max, J. Quantizing for Minimum Distortion. IRE Trans. Inf. Theory 1960, 6, 7–12. [Google Scholar] [CrossRef]
  40. Stuart, P. Lloyd, Least squares quantization in PCM. IEEE Trans. Inf. Theory 1982, 28, 129–136. Available online: https://www.doc88.com/p-6631815894116.html (accessed on 1 January 2024).
  41. Mun, S.; Fowler, J.E. Block compressed sensing of images using directional transforms. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 3021–3024. [Google Scholar] [CrossRef]
  42. Sun, L.; Hays, J. Super-resolution from internet-scale scene matching. In Proceedings of the 2012 IEEE International Conference on Computational Photography (ICCP), Seattle, WA, USA, 28–29 April 2012; pp. 1–12. [Google Scholar]
  43. Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
Figure 1. The CS coding structure based on the proposed method.
Figure 1. The CS coding structure based on the proposed method.
Entropy 26 00468 g001
Figure 2. The curve and gradient curve of the proposed activation function.
Figure 2. The curve and gradient curve of the proposed activation function.
Entropy 26 00468 g002
Figure 3. Network structure diagram of CNN based on model (5).
Figure 3. Network structure diagram of CNN based on model (5).
Entropy 26 00468 g003
Figure 4. Structure diagram of the proposed BCS measurement quantization process.
Figure 4. Structure diagram of the proposed BCS measurement quantization process.
Entropy 26 00468 g004
Figure 5. Network structure diagram for estimating the location parameters.
Figure 5. Network structure diagram for estimating the location parameters.
Entropy 26 00468 g005
Figure 6. Network structure diagram for estimating the scale parameters.
Figure 6. Network structure diagram for estimating the scale parameters.
Entropy 26 00468 g006
Figure 7. Structure diagram of the proposed dequantization process.
Figure 7. Structure diagram of the proposed dequantization process.
Entropy 26 00468 g007
Figure 8. Network structure diagram of the inverse process of the proposed quantization.
Figure 8. Network structure diagram of the inverse process of the proposed quantization.
Entropy 26 00468 g008
Figure 9. Network structure diagram of the measurement information correction module.
Figure 9. Network structure diagram of the measurement information correction module.
Entropy 26 00468 g009
Figure 10. Network structure diagram for quantizing parameters.
Figure 10. Network structure diagram for quantizing parameters.
Entropy 26 00468 g010
Figure 11. Network structure diagram for dequantizing parameters.
Figure 11. Network structure diagram for dequantizing parameters.
Entropy 26 00468 g011
Figure 12. Eight test images.
Figure 12. Eight test images.
Entropy 26 00468 g012
Figure 13. PSNR curves of the eight test images.
Figure 13. PSNR curves of the eight test images.
Entropy 26 00468 g013aEntropy 26 00468 g013b
Figure 14. Visual comparison of different methods at a compression bit rate of 0.1.
Figure 14. Visual comparison of different methods at a compression bit rate of 0.1.
Entropy 26 00468 g014aEntropy 26 00468 g014b
Figure 15. Average PSNR curves of test image sets.
Figure 15. Average PSNR curves of test image sets.
Entropy 26 00468 g015
Table 1. Number of reconstruction levels for scalar quantization methods.
Table 1. Number of reconstruction levels for scalar quantization methods.
Quantization Bit-Depth2345678
Number of reconstruction levels48163264128256
Table 2. Number of reconstruction levels for the proposed quantization.
Table 2. Number of reconstruction levels for the proposed quantization.
Quantization Bit-Depth2345678
APC147236397690114718602921
Aerial199321559981171228634368
Airplane181273453724112216442336
Airport186283488845148325454009
Building161243414739130222773699
Moon surface171270455782132822053515
Tank161241413744134524223963
Truck144227396699122021403527
Average168.7261.7446.8775.51332.32244.53542.2
Table 3. MSE of the measurements quantized with 3-bit.
Table 3. MSE of the measurements quantized with 3-bit.
ImageUniform Quantizationµ-Law QuantizationLloyd–MaxProposed Method
APC567.30675.16493.99115.76
Aerial1468.861250.971223.14192.34
Airplane1147.161280.451279.69161.52
Airport704.45435.69460.30155.19
Building864.87768.95629.43113.59
Moon surface1327.97831.08573.44115.48
Tank780.74730.11634.44121.87
Truck540.80489.04486.02121.88
Average925.27807.68722.56137.20
Table 4. MSE of the measurements quantized with 6-bit.
Table 4. MSE of the measurements quantized with 6-bit.
ImageUniform Quantizationµ-Law QuantizationLloyd–MaxProposed Method
APC6.867.1511.432.53
Aerial17.7716.1121.373.87
Airplane14.0413.8619.623.20
Airport8.275.419.033.33
Building11.029.7812.892.51
Moon surface15.1110.6811.662.58
Tank9.618.5812.322.63
Truck6.605.8810.212.63
Average11.169.6813.572.91
Table 5. MSE of the measurements quantized with 8-bit.
Table 5. MSE of the measurements quantized with 8-bit.
ImageUniform Quantizationµ-Law QuantizationLloyd–MaxProposed Method
APC0.41940.43371.87160.1742
Aerial1.06921.01552.66580.2978
Airplane0.85460.83042.54700.2356
Airport0.50310.32741.58680.2543
Building0.66840.59272.02410.1770
Moon surface0.93080.64401.89950.1778
Tank0.58840.51041.91210.1980
Truck0.40320.36301.72780.1902
Average0.67960.58962.02930.2131
Table 6. MSE and entropy for different λ .
Table 6. MSE and entropy for different λ .
λ MSEEntropy
00.28837.2624
0.050.28417.3376
0.10.28677.3539
0.20.28967.3787
0.50.30277.4487
Table 7. Quantization performance of different information correction modules.
Table 7. Quantization performance of different information correction modules.
MSEEntropy
µ-law quantification1.09417.1510
Contrast Scheme 10.32987.1760
Contrast Scheme 20.30277.2430
Contrast Scheme 30.28417.3376
Table 8. The PSNRs of the four datasets at a bit rate of 0.1 bpp.
Table 8. The PSNRs of the four datasets at a bit rate of 0.1 bpp.
Set5Set11Set14BSD68SunHays80Urban100Average
Uniform Quantization + Entropy CodingPSNR19.5918.2118.8119.4919.3220.0119.24
SSIM0.21170.15260.17690.15470.13610.21170.1739
µ-law QuantizationPSNR17.7916.1716.7217.4017.3019.8617.54
SSIM0.15460.11760.13360.10800.10010.23110.1408
Lloyd–Max QuantizationPSNR19.1217.7918.2718.9718.7519.1218.67
SSIM0.19090.13550.15870.13280.11960.19090.1547
Proposed MethodPSNR20.2218.6119.1720.0019.9020.2519.69
SSIM0.22290.16720.18750.16360.14890.22290.1855
Table 9. The PSNRs of the four datasets at a bit rate of 0.2 bpp.
Table 9. The PSNRs of the four datasets at a bit rate of 0.2 bpp.
Set5Set11Set14BSD68SunHays80Urban100Average
Uniform Quantization + Entropy CodingPSNR22.2619.4120.3221.2720.9922.2621.09
SSIM0.32460.22690.26530.24210.22600.32460.2683
µ-law QuantizationPSNR22.0019.2820.1521.0520.8421.9420.88
SSIM0.31270.21000.25160.22750.20960.31410.2543
Lloyd–Max QuantizationPSNR22.0319.3420.2221.1120.8722.0320.93
SSIM0.31270.21680.25420.22980.21390.31270.2567
Proposed MethodPSNR22.4319.670720.5521.4221.1322.4321.27
SSIM0.32490.24210.27250.24730.23130.32490.2738
Table 10. Detailed network structures of parameter estimation.
Table 10. Detailed network structures of parameter estimation.
Convolution LayerKernel SizeStrideInput ChannelsOutput ChannelsActivation Function
Conv110 × 1113LeakyReLU
Conv21 × 1133LeakyReLU
Conv31 × l131-
Table 11. Detailed network structures of the quantization process.
Table 11. Detailed network structures of the quantization process.
Convolution LayerKernel SizeStrideInput ChannelsOutput ChannelsActivation Function
Conv11 × 11l6LeakyReLU
Conv21 × 1166LeakyReLU
Conv31 × 1161 g 3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gong, J.; Chen, Q.; Zhu, W.; Wang, Z. A Convolutional Neural Network-Based Quantization Method for Block Compressed Sensing of Images. Entropy 2024, 26, 468. https://doi.org/10.3390/e26060468

AMA Style

Gong J, Chen Q, Zhu W, Wang Z. A Convolutional Neural Network-Based Quantization Method for Block Compressed Sensing of Images. Entropy. 2024; 26(6):468. https://doi.org/10.3390/e26060468

Chicago/Turabian Style

Gong, Jiulu, Qunlin Chen, Wei Zhu, and Zepeng Wang. 2024. "A Convolutional Neural Network-Based Quantization Method for Block Compressed Sensing of Images" Entropy 26, no. 6: 468. https://doi.org/10.3390/e26060468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop