Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Dynamic Divide Grouping Non-Orthogonal Multiple Access in Terrestrial-Satellite Integrated Network
Previous Article in Journal
Autofocus Entropy Repositioning Method Bioinspired in the Magnetic Field Memory of the Bees Applied to Pollination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual Optical Path Based Adaptive Compressive Sensing Imaging System

1
School of Engineering Science, University of Chinese Academy of Sciences, Beijing 100049, China
2
Institute of Computing Technology, CAS, Beijing 100190, China
3
School of Information Science and Technology, University of Science and Technology of China, Hefei 230027, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(18), 6200; https://doi.org/10.3390/s21186200
Submission received: 17 August 2021 / Revised: 8 September 2021 / Accepted: 10 September 2021 / Published: 16 September 2021
(This article belongs to the Section Optical Sensors)

Abstract

:
Compressive Sensing (CS) has proved to be an effective theory in the field of image acquisition. However, in order to distinguish the difference between the measurement matrices, the CS imaging system needs to have a higher signal sampling accuracy. At the same time, affected by the noise of the light path and the circuit, the measurements finally obtained are noisy, which directly affects the imaging quality. We propose a dual-optical imaging system that uses the bidirectional reflection characteristics of digital micromirror devices (DMD) to simultaneously acquire CS measurements and images under the same viewing angle. Since deep neural networks have powerful modeling capabilities, we trained the filter network and the reconstruction network separately. The filter network is used to filter the noise in the measurements, and the reconstruction network is used to reconstruct the CS image. Experiments have proved that the method we proposed can filter the noise in the sampling process of the CS system, and can significantly improve the quality of image reconstruction under a variety of algorithms.

1. Introduction

With the development of mobile Internet and imaging technology, a large number of images are generated in daily life. As the resolution of the images keeps increasing, the density of the photoelectric sensor will be higher and higher. However, the signal acquisition needs to meet the classic Shannon–Nyquist sampling theorem, which means that, if you want to restore the sampled signal without distortion, the sampling frequency needs to be greater than twice the highest frequency in the signal [1]. For this reason, high-resolution image sensors will generate a large amount of data when taking images. Normally, we will compress these image data with some compression formats, such as the well-known JPEG and PNG. In this process, a large amount of redundant data will be removed, and which will also consume a lot of computing resources. For this reason, we urgently need a method that can directly sample the compressed data. Compressive Sensing [2] is just such a theory that can sample a small amount of compressed data to reconstruct high-quality images.
Compared with the traditional sampling compression theory, the biggest advantage of CS is that the sampling and compression processes are carried out at the same time, and the compressed data can be collected directly. However, the theory assumes that the sampled signal needs to meet a certain sparsity. The proposal of this theory enables signal acquisition to break through the Shannon–Nyquist sampling theorem under certain conditions, and can restore the original signal by sampling a small amount of data, which makes the theory have a wide range of applications in practice, such as medical imaging, remote sensing imaging, and sensor networks.
In the practical application of compressive sensing, a spatial light modulator is used to obtain the measurements of the image by displaying the measurement matrix. However, as the difference between the rows of the measurement matrix is very small, the brightness of the light after being modulated by the spatial light modulator is actually similar, which requires a high-sensitivity photoelectric sensor to sample. However, this sensor will fluctuate greatly in response to small changes of ambient light. At the same time, we usually need to perform a high gain of the electric signal in the imaging system. For example, in our hardware device, mentioned later in this paper, the gain is up to 10 8 times. Such a large gain will cause the noise signal to be amplified many times, and even drown the strength of the original signal. In addition, in order to improve the collection speed of the equipment, it is usually necessary to switch the measurement matrix image at a high rate with the spatial light modulator. This requires a high synchronization between the photodiode and the spatial light modulator. The small differences will cause the photodiode to fail to collect the right light signal. In summary, the imaging quality of current CS imaging systems is very poor.
The imaging quality of several imaging systems is shown in Figure 1. Among them, Figure 1a shows the image quality of the single-pixel camera (SPC) [3]. Although there is only a simple letter, and the image has only black and white colors, the edges of the image are blurred. Figure 1b is the imaging quality of Reconnect, and Figure 1c is the result of our imaging system with a 25% measurement rate. We can see that the image quality is very poor. This image quality directly affects the application of CS imaging systems. For this reason, we urgently need a CS imaging system that can adapt to its own noise.
For this reason, we propose an adaptive CS imaging system based on a dual optical path. The main contributions are summarized as follows:
(1)
Aiming at the problem that deep learning data cannot be labeled in the actual CS sampling, a dual light path acquisition and labeling method is proposed, and this method is used to obtain the data required for network training;
(2)
A filtering method for a CS imaging system based on deep learning is proposed. At the same time, a reconstruction network is constructed on the basis of this filtering method to better reconstruct CS measurement data;
(3)
A CS imaging system is established to adaptively filter the noise created by the hardware, which can significantly improve the imaging quality so that the CS imaging device can be better applied to image acquisition.
Experiments show that this filtering method not only has good reconstruction results based on deep learning reconstruction methods, but can also improve the reconstruction quality of other traditional algorithms.
The organization of this paper is as follows: Section 2 briefly introduces the basic concepts of CS and existing imaging systems. In Section 3, the dual optical path acquisition system and the reconstruction algorithm are proposed in detail. The experimental results are shown and analyzed in Section 4. Finally, the conclusion of this paper is presented in Section 5.

2. Basic Concepts and Related Work

The related work will be reviewed from the perspective of the basic principles of CS imaging and some imaging systems.

2.1. Compressive Sensing

Compressive Sensing is a theory of signal compression acquisition. Its basic principle is to project sparse signals into a low-dimensional space through a special measurement matrix for acquisition. This model was proposed by Donoho, Candes and Tao in their work [2] in 2006. The basic mathematical model is expressed as follows: Assuming the signal to be compressed is x R N , then the sampling process can be defined as the following equation:
y = Φ x ,
where Φ R M × N is called the measurement matrix, which is to project the original signal into a low-dimensional space. y R M is the compressed signal, called the measurements. The dimension of measurements is much smaller than the dimension of the original signal, which is M N .
The reconstruction of CS is a process of recovering an approximate solution of the signal x through a certain algorithm using the measurements y . As we know, the dimension of y is much smaller than the dimension of x . Therefore, the solution from y to x is an NP-hard problem [4]. For practical problems, it is usually necessary to introduce other assumptions to solve this problem.
Traditional CS imaging reconstruction algorithms usually consider the structural characteristics of the image, such as the total variation (TV) model. The minimum total variation model assumes that the image has sparseness in vertical and horizontal gradients, so this method is widely used in CS image reconstruction [5,6]. For the solution of this method, TVAL3 [7] is a recently proposed TV-based CS image reconstruction algorithm. This algorithm not only has a fast convergence speed, but also supports a variety of measurement matrices. It is a commonly used algorithm that can recover image details. The reconstruction algorithm based on the image structure characteristics also includes the NLR-CS [8] method. This method is based on the similar characteristics of non-local images and reconstructs CS images according to the low-rank characteristics of non-local image blocks. This algorithm is also a reconstruction algorithm based on image structure characteristics. There are also some algorithms based on image filters, among which D-AMP, proposed by [9], is more widely used.
In recent years, some scholars have used neural networks to solve the problem of CS reconstruction. DNN was first introduced into CS recovery by Mousavi in the work on the stacked denoising autoencoder (SDA) [10]. The work proposed an autoencoder based on multi-layer perceptions (MLPs) to solve the CS recovery problem. Inspired by the work on super-resolution (SRCNN) [11], Kulkarni et al. proposed a CNN-based neural network called ReconNet in their work [1]. This work first introduced convolution neural networks (CNN) to CS and the performance is better than that of SDA. However, these methods do not consider the structural characteristics of the image, such as the saliency of the image. Compressed sensing based on blocks can more effectively process natural images [12,13]. These methods divide the image into non-overlapping blocks and perform separate measurements on different image blocks. In [14], fewer measurement rates are allocated to non-salient image blocks but more are allocated to salient blocks. This can effectively improve the efficiency of compressive sensing measurements, thereby improving the reconstruction quality of the saliency area. However, this method will cause obvious blocking artifacts. In [15], a multi-channel deep network was proposed for a full image to remove blocking artifacts. AMP-Net [16] iteratively uses denoising and deblocking methods to eliminate block artifacts in the image. Since the actual light cannot be accurately divided into blocks, these methods cannot be applied to actual compressive sensing image acquisition.
There are also several works based on the recent, popular Generative Adversarial Networks (GAN). The work [17] added a discriminator to the network of ReconNet. Dave et al. implemented the generative model RIDE [18] for pixel prediction to recover the compressive image in [19]. This work reconstructed the image pixel by pixel, which is a new type of reconstruction method. In [17], a fully connected layer was added to the ReconNet architecture to learn the measurements from the scene image. The work in [20] used the method of a joint learning measurement matrix, which greatly improves the reconstruction quality. These methods of learning measurement matrices can obtain matrices that are more suitable for natural images, but the training process for these matrices is complicated and at the same time highly dependent on the training set. In recent years, DNN has also been applied to CS video recovery [21,22].

2.2. Imaging System

The basic theory of CS sampling is y = Φ x , where x is the light in the scene we need to sample and Φ is the measurement matrix. The key problem is how to perform matrix multiplication on the light and how to obtain the measurements. We know that the spatial light modulator can control the amount of light passing through different pixel positions to achieve modulation of the light, such as the commonly used devices: liquid crystal panels, digital micromirror devices (DMD), and so forth. The modulated light is received by the photoelectric sensor to obtain the measurements.
Using DMD to modulate light, Duarte proposed a CS imaging system: the Single Pixel Camera (SPC) [3]. The light reflected by the objects in the scene is imaged onto the DMD through the convergence of the lens. Then, the DMD modulates the measurement matrix to reflect the modulated light afterwards. Finally, the modulated light is sampled by the photodiode. SPC is the first acquisition system for CS imaging. The system uses one pixel to reconstruct an image with a resolution of 256 × 256 . However, due to the acquisition noise, its imaging quality is poor.
Using the LCD panel as the spatial light modulator, Huang proposed a lensless camera [23]. Using the principle of small aperture imaging, the camera does not require a lens. However, as we know, the lens is used in the traditional camera to converge the light and increase the imaging field of view. As the light transmittance of the liquid crystal is poor, most of the light cannot pass through the liquid crystal, which leads to poor image reconstruction quality.
The above two cameras are based on the traditional camera model, which receives the light reflected by the environment and modulates ir. Some scholars have also proposed an imaging system based on active light. Zhang [24] proposed a system that uses a projector to perform CS imaging. The camera uses a projector to project the measurement matrix into the scene to be sampled. However, this kind of camera collects the light of the entire projection space, which means that other light in the environment will affect the system and increase the noise of the reconstructed image.
At present, CS imaging systems require high sampling accuracy, and generally require high-precision signal amplification and filtering circuits. However, due to the influence of ambient light and circuit noise, the quality of image reconstruction is poor. Therefore, a CS imaging system that can reconstruct images with high quality is needed.

3. Adaptive Compressive Sensing Imaging System

In recent years, with the development of deep neural networks, many impressive results have been achieved in the field of image processing. The advantage of neural networks is that they can simulate complex mathematical models, but their disadvantage is that they are data-driven and require a lot of training data. Based on this feature, we propose a system to use neural networks to remove noise in CS imaging hardware and improve imaging quality. Its basic structure is shown in Figure 2. In order to obtain the data needed for neural network training, we designed a data acquisition system based on dual optical paths. Next, we will give a detailed introduction of the basic structure and hardware system of the dual optical path design.

3.1. Dual Optical Path Sampling

The purpose of the dual-optical acquisition system is to obtain image data and CS sampling data of the same scene at the same time, so as to calibrate the measurements of CS and optimize the reconstruction results. To achieve this purpose, we use the bidirectional reflection characteristics of DMD. In other words, the DMD micromirror can reflect light at ± 12 degrees. Usually in CS imaging hardware, DMD is used to modulate the measurement matrix and the scene light. However, only + 12 degrees light is used in this process. In fact, the 12 degrees optical path can also obtain the image information. More importantly, the viewing angle of the scene observed by the ± 12 degrees optical path is the same, which provides a guarantee for us to design a dual optical path acquisition and calibration system.
As is shown in Figure 3, the light in the scene is converged by the viewfinder lens and is imaged on DMD. At this time, the image is divided into discrete image pixels by micromirrors on the DMD, where each mirror represents a pixel. The forward-reflected light is converged on the photoelectric sensor through the CS Sampling optical path lens. At this time, if each row of the measurement matrix is displayed one by one on the DMD, we can obtain the measurements y of CS. The reverse light is imaged on the CMOS image sensor through the CMOS light path lens. At this time, if the DMD displays a white image, then the CMOS will capture all the light in the scene, which is the scene image x . In this way, we will obtain a set of label data < x , y >, where y is a set of measurements with the optical path and circuit noise in the CS acquisition hardware, and x is the target image data. At this point, we have obtained the calibration data required for the training of the CS imaging system.
As shown in Figure 4, in order to avoid the influence of the circuit on the optical path, the hardware of the sampling system is divided into two parts: the optical sampling part and the circuit control part.These two parts are shielded from light. In the optical sampling part, the main lens is used to control the aperture, focal length and focus, and so forth, for the convenient sampling of different scenes. DMD uses the D4100 produced by Texas Instruments (TI) with a pixel resolution of 1920 × 1080 . Among them, the CMOS cannot completely align the image on the DMD, and we also only use some of the pixels in the center of the DMD, so the image obtained by the CMOS needs to be calibrated before it can be used. The circuit control part mainly includes the controller of each sampling part and a signal amplifier, among which the amplification factor of the signal amplifier reaches 10 8 times.

3.2. Recovery Network

To adaptively filter the device noise, we need to use the characteristics of neural networks to train the noise model of the imaging device. After this, the filtered measurements are obtained, which are then reconstructed to improve the imaging quality of the device. For this reason, the design of the neural network is divided into a filter network and a reconstruction network.

3.2.1. Filter Network

Considering that the poor imaging quality of the CS imaging system is mainly because the sampled measurements contain a lot of noise, we first build a CNN network to denoise the measurements. As shown in Figure 5, a 4-layer denoising network is built based on convolution. This process can be expressed as y ^ = C ( y , W c ) , where y ^ is the filtered measurements, and W c is the weights of the convolution layer. The basic structure of the network is shown in Figure 5, where each convolution layer is followed by batch normalization and the ReLU activation function. The first convolution layer takes a kernel size of 11 and outputs 32 feature maps. The second takes a kernel size of 1 and outputs 64 feature maps. The third takes a kernel size of 7 and outputs 32 feature maps. The last layer takes a kernel size of 3 and generates 1 × M measurements. To obtain W c , MSE is employed as the loss function.

3.2.2. Reconstruction Network

The measurements obtained through the filter network can already be used in the reconstruction in traditional CS algorithms, but in order to obtain better reconstruction quality, we ran a reconstruction network after the filter network. The network is composed of two fully connected layers. Assuming that the approximate value of the image output by the network is x ^ , then x ^ = L ( y , W l ) , where W l is the weights of the fully connected layers, which are trained by another MSE loss function.

3.2.3. Training

The training of the neural network of the aforementioned imaging system is divided into three parts; the first part is the training of the filter network, the second part is the training of the reconstruction network, and the third part is the joint training. We set the sample image size to 128 × 128 ; a total of 44 images were sampled on the device, of which eight images were used as the test set.

Filter Network Training

Since the measurements are a 1 × 16,384 signal, we intercept the measurements with a length of M. The specific value of M is related to the sampling rate. The lengths of 1 % , 4 % , 10 % , and 25 % are 164, 655, 1638, 4096, respectively. After setting the step size to 20, we obtained about 27,000 sets of training data.

Reconstruction Network Training

The training of this network is similar to the general training of a CS reconstruction network based on deep learning, such as ReconNet. Use the existing image dataset, we simulate the measurements and then train the network. Here we choose about 14,000 sets of training data.

Joint Training

Joint training refers to the combination of the above two networks, the initial loading of the two parts of the network weights that have been pre-trained, reducing the learning rate of the network, and training on 36 datasets sampled by our device. All network training is implemented on pytorch.
The proposed network is implemented on the public Pytorch framework. The weights of the networks are initialized by normal Gaussian distribution with the stand variance of 0.01. The optimizer algorithm for all training is Adam. We run the training model on a PC with 2.1 GHz CPU, 128 GB RAM and a NVIDIA GeForce RTX 2080Ti GPU. The learning rate is initialized with 1 × 10 4 and decreased by 50 % at 20 epochs. The whole training procedure takes 100 epochs for each network. The training time for each method takes about 8 h.

4. Experimental Results

4.1. Comparison with Existing Methods

In order to verify the imaging effect of our reconstruction network, we compare the reconstruction results of different algorithms at the same measurement rate (MR) and the same measurements sampled by our hardware device. The algorithms include TVAL3 [7], NLR-CS [8] and D-AMP [9]. The PSNR and SSIM results of eight images are shown in Table 1. The reconstruction images of a 25% measurement rate can be found in Figure 6. The quantization of all images uses 8 bits, including the calculation of reconstructed images and PSNR. As the sampled measurements are noisy, some algorithms cannot even reconstruct images, such as partial examples of TVAL3 and NLR-CS. It can be seen that our algorithm has the best reconstruction quality.
Compared with the existing methods, our method has the following advantages:
  • It can be seen from the average value in Table 1 that our method achieves the best reconstruction performance. Our method achieves the best PSNR and SSIM under all the four measurement rates; especially at high sampling rates, the PSNR and SSIM are much higher;
  • Observing the results, we can find that the result of the D-AMP algorithm is the second best, and it is obviously better than the other two algorithms. The main reason is that the D-AMP algorithm is an iterative algorithm based on a filter, and the filter has a certain effect on the noise, but the effect is not very good;
  • The results of TVAL3, NLR-CS and D-AMP at a high measurement rate of 0.25 are not better than those at a low measurement rate of 0.10. The main reason is that the device samples more data at high measurement rates and the acquisition time is long, which will introduce more cumulative errors.
Under the same environment, the average reconstruction time of TVAL3, NLR-CS and D-AMP for each image is 0.23, 30.11, and 8.25 s, respectively. Since the neural network can be accelerated by the GPU, the average reconstruction time of the reconstruction network is 0.02 s. When only the CPU is considered for reconstruction, the average obtained reconstruction time is 0.11 s.
From Figure 6, we can see that the unfiltered data reconstruction result will have some vertical noise in the fixed area of the image, which is present in almost all images. Compared with other unfiltered algorithms, the reconstruction results of D-AMP are a little better, but there is a serious phenomenon of denoising and smearing, which leads to blurry images. Our method filters out noise while preserving details.

4.2. Evaluation of Filter Network

To evaluate the performance of the filter network, we first pass the measurements through our filter network, and then reconstruct it with the TVAL3, NLR-CS and D-AMP algorithms again. The results obtained are shown in Table 2. The reconstruction result is the average PSNR and SSIM of eight test images. Comparing with the previous results, we can find the following characteristics:
  • The quality of the reconstruction results of TVAL3, NLR-CS and D-AMP has been significantly improved. The increase in PSNR of some algorithms is about 7 dB;
  • The higher the measurement rate, the greater the improvement in image quality. The main reason is that the measurement rate is high, the device acquisition time is long, and the cumulative error introduced will be large, so the effect achieved by the filter network is better;
  • In fact, the reconstruction result of 0.01 is already very poor; the image basically cannot show the content of the original image. The SSIM of the image is also very low and the PSNR of D-AMP after filtering is reduced, which is meaningless.
At the same time, we compare the difference between the sampled signal, the filtered signal and the original signal. It is obvious that a low-frequency noise signal is added onto the sampled signal, which actually includes some high-frequency signal noise. After the process of our filter network, the signal is basically the same as the original signal, so the reconstruction quality is significantly improved. By comparison, the error between the filtered signal and the original signal is much smaller, indicating that the filter network has a significant effect on the noise of the CS imaging device.
The advantage of the compressive sensing imaging system proposed in this work is the use of dual optical paths for sampling. The system can not only provide the measurements of compressive sensing imaging, but can also calibrate the measurements so that no matter which reconstruction method is used, a better reconstruction quality can be obtained. Experiments have proved that the compressive sensing sampling method based on dual optical paths can achieve better image quality in a variety of reconstruction methods. As shown in Figure 7a, there is obviously a low-frequency noise in the sampled signal. The signal after filtering obviously removes the low-frequency noise signal and is very similar to the original signal. As shown in Figure 7b, the mean square error of the filtered signal is greatly reduced for each image.

5. Conclusions and Future Work

Because CS imaging systems have high sensitivity to ambient light and circuit noise, even if we add a variety of hardware filtering methods, their imaging quality is still very poor. To this end, we propose a dual optical path CS imaging system. The design of the dual optical path is based on the bidirectional reflection characteristics of DMD, so that one optical signal is used for CS measurements sampling, and the other optical signal is used for CMOS images acquisition. In this way, we labeled the measurements with the images in the same viewing angle of the scene, and this labeled data are required for neural network training. We designed and implemented a dual optical path imaging system, and also proposed a filter and reconstruction network. After training, this reconstruction network can obtain better imaging results. Through the filter network, the noise in the measurements is suppressed, and the reconstruction quality is significantly improved even for the traditional CS reconstruction algorithms. Experiments have verified that this dual optical path based CS imaging system can effectively improve the imaging quality and suppress the noise caused by the optical path and the circuit on the measurements. The filter network can adaptively filter the noise created by the hardware, so that the CS imaging device can be better applied to image acquisition.
Although the method we proposed can improve the imaging quality of a CS imaging system, the system will take a large amount of image data during its work. Therefore, future work will focus on another technique that is required to use these newly generated image data for training and to continuously update the filter network online.

Author Contributions

All authors equally contributed to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China grant number (62032022, 61972375, 61871258, 61671426) and the Fundamental Research Funds for the Central Universities grant number E0E48980.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kulkarni, K.; Lohit, S.; Turaga, P.; Kerviche, R.; Ashok, A. Reconnet: Non-iterative reconstruction of images from compressively sensed measurements. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 30 June 2016; pp. 449–458. [Google Scholar]
  2. Candes, E.J.; Romberg, J.; Tao, T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar] [CrossRef] [Green Version]
  3. Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Sun, T.; Kelly, K.F.; Baraniuk, R.G. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef] [Green Version]
  4. Baraniuk, R.G. Compressive Sensing [Lecture Notes]. IEEE Signal Process. Mag. 2007, 24, 118–121. [Google Scholar] [CrossRef]
  5. Antonin, C.; Pierre-Louis, L. Image recovery via total variation minimization and related problems. Numer. Math. 1997, 76, 167–188. [Google Scholar]
  6. Tony, C.; Selim, E.; Frederick, P.; Yip, A. Recent developments in total variation image restoration. Math. Model. Comput. Vis. 2005, 17, 2. [Google Scholar]
  7. Li, C.; Yin, W.; Jiang, H.; Zhang, Y. An efficient augmented lagrangian method with applications to total variation minimization. CAAM Technical Report. Comput. Optim. Appl. 2013, 56, 507–530. [Google Scholar] [CrossRef] [Green Version]
  8. Dong, W.; Shi, G.; Li, X.; Ma, Y.; Huang, F. Compressive sensing via nonlocal low-rank regularization. IEEE Trans. Image Process. Publ. 2014, 23, 3618–3632. [Google Scholar] [CrossRef] [PubMed]
  9. Metzler, C.A.; Maleki, A.; Baraniuk, R.G. From denoising to compressed sensing. IEEE Trans. Inf. Theory 2016, 62, 5117–5144. [Google Scholar] [CrossRef]
  10. Mousavi, A.; Patel, A.B.; Baraniuk, R.G. A deep learning approach to structured signal recovery. Communication, Control, and Computing (Allerton). In Proceedings of the 2015 53rd Annual Allerton Conference, Monticello, IL, USA, 29 September 2015; pp. 1336–1343. [Google Scholar]
  11. Dong, C.; Loy, C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar]
  12. Adler, A.; Boublil, D.; Zibulevsky, M. Block-based compressed sensing of images via deep learning. In Proceedings of the 2017 IEEE 19th International Workshop on Multimedia Signal Processing (MMSP), London, UK, 16–18 October 2017; pp. 1–6. [Google Scholar]
  13. Akbari, A.; Trocan, M. Robust image reconstruction for block-based compressed sensing using a binary measurement matrix. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 1832–1836. [Google Scholar]
  14. Yu, Y.; Wang, B.; Zhang, L. Saliency-based compressive sampling for image signals. IEEE Signal Process. Lett. 2010, 17, 973–976. [Google Scholar]
  15. Zhou, S.; He, Y.; Liu, Y. Multi-Channel Deep Networks for Block-Based Image Compressive Sensing. IEEE Trans. Multimed. 2021, 23, 2627–2640. [Google Scholar] [CrossRef]
  16. Zhang, Z.; Liu, Y.; Liu, J. AMP-Net: Denoising-based deep unfolding for compressive image sensing. IEEE Trans. Image Process. 2020, 30, 1487–1500. [Google Scholar] [CrossRef] [PubMed]
  17. Lohit, S.; Kulkarni, K.; Kerviche, R.; Turaga, P.K.; Ashok, A. Convolutional Neural Networks for Noniterative Reconstruction of Compressively Sensed Images. IEEE Trans. Comput. Imaging 2018, 4, 326–340. [Google Scholar] [CrossRef] [Green Version]
  18. Theis, L.; Bethge, M. Generative image modeling using spatial LSTMs. Adv. Neural Inf. Process. Syst. 2015, 28, 1927–1935. [Google Scholar]
  19. Dave, A.; Vadathya, A.K.; Mitra, K. Compressive image recovery using recurrent generative model. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 1702–1706. [Google Scholar]
  20. Mdrafi, R.; Gurbuz, A.C. Joint learning of measurement matrix and signal reconstruction via deep learning. IEEE Trans. Comput. Imaging 2020, 6, 818–829. [Google Scholar] [CrossRef]
  21. Mu, Q.; Meng, Z.; Ma, J.; Yuan, X. Deep learning for video compressive sensing. Apl Photonics 2020, 5, 030801. [Google Scholar]
  22. Cheng, Z.; Lu, R.; Wang, Z.; Zhang, H.; Chen, B.; Meng, Z.; Yuan, X. BIRNAT: Bidirectional recurrent neural networks with adversarial training for video snapshot compressive imaging. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 258–275. [Google Scholar]
  23. Huang, G.; Jiang, H.; Matthews, K.; Wilford, P. Lensless imaging by compressive sensing. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Melbourne, Australia, 15–18 September 2013; pp. 2101–2105. [Google Scholar]
  24. Zhang, Z.; Ma, X.; Zhong, J. Single-pixel imaging by means of Fourier spectrum acquisition. Nat. Commun. 2015, 6, 6225. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Imaging quality of several systems. (a) Single pixel camera; (b) ReconNet hardware imaging; (c) Our hardware system imaging.
Figure 1. Imaging quality of several systems. (a) Single pixel camera; (b) ReconNet hardware imaging; (c) Our hardware system imaging.
Sensors 21 06200 g001
Figure 2. Architecture of dual optical path based adaptive CS imaging system.
Figure 2. Architecture of dual optical path based adaptive CS imaging system.
Sensors 21 06200 g002
Figure 3. Dual optical path design. (a) Optical path design; (b) Hardware design.
Figure 3. Dual optical path design. (a) Optical path design; (b) Hardware design.
Sensors 21 06200 g003
Figure 4. Sampling system architecture.
Figure 4. Sampling system architecture.
Sensors 21 06200 g004
Figure 5. Neural networks of the imaging system.
Figure 5. Neural networks of the imaging system.
Sensors 21 06200 g005
Figure 6. Reconstruction results of the images.
Figure 6. Reconstruction results of the images.
Sensors 21 06200 g006
Figure 7. The effect of the filter network on the signal. (a) Raw data of sampled signal and filtered signal; (b) The mean square error of the sampled signal and the filtered signal.
Figure 7. The effect of the filter network on the signal. (a) Raw data of sampled signal and filtered signal; (b) The mean square error of the sampled signal and the filtered signal.
Sensors 21 06200 g007
Table 1. The reconstruction results of different algorithms at different MR. The inputs of the algorithms are the data directly sampled by hardware. The best results are in bold. Our algorithm can achieve the best results at each MR.
Table 1. The reconstruction results of different algorithms at different MR. The inputs of the algorithms are the data directly sampled by hardware. The best results are in bold. Our algorithm can achieve the best results at each MR.
ImageAlgorithmMR = 0.25MR = 0.10MR = 0.04MR = 0.01
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
BarbaraTVAL39.840.4310.230.4211.660.0312.510.23
NLR-CS9.790.4210.640.399.990.3310.960.28
D-AMP12.800.5812.180.4911.880.4014.170.32
Ours22.650.7022.610.7720.830.6814.010.38
BoatsTVAL39.950.448.320.387.450.306.470.20
NLR-CS9.570.417.640.327.290.286.050.17
D-AMP11.980.5011.420.4110.440.3712.660.29
Ours20.760.6419.970.7117.750.6315.920.43
CameramanTVAL39.680.469.850.4811.060.4410.010.30
NLR-CS9.710.4310.200.4511.120.4010.020.30
D-AMP12.860.5512.490.5512.500.4611.030.32
Ours18.800.4118.100.5017.340.3014.390.18
ParrotsTVAL310.380.0.539.530.419.850.419.600.23
NLR-CS10.130.509.380.369.350.359.460.20
D-AMP11.770.5413.320.5811.910.5010.570.27
Ours23.270.7022.860.8018.930.6115.920.52
FingerprintTVAL310.540.2112.570.2110.110.116.600.10
NLR-CS10.340.2012.200.218.080.156.770.10
D-AMP12.760.2312.460.2210.630.1615.330.17
Ours17.790.3018.140.3417.280.2914.960.20
ForemanTVAL33.440.0612.110.473.460.064.360.19
NLR-CS3.450.065.250.294.860.244.320.17
D-AMP11.400.5613.440.5913.030.5212.400.47
Ours20.990.7717.930.7720.930.7714.580.36
HouseTVAL34.230.0912.430.434.450.125.450.21
NLR-CS4.300.105.860.285.150.215.420.19
D-AMP10.990.5411.500.4611.530.4912.940.44
Ours24.420.7217.570.8217.590.7016.870.42
LenaTVAL311.810.4814.090.2511.990.2511.790.28
NLR-CS11.380.4610.960.4010.920.3711.730.26
D-AMP12.320.5512.010.4511.560.4210.040.21
Ours22.140.7622.180.8720.510.7614.520.28
MeanTVAL38.730.3411.140.388.750.228.340.22
NLR-CS8.580.329.020.348.350.298.090.21
D-AMP12.110.5112.350.4711.690.4212.390.31
Ours21.350.6319.920.7018.890.5915.140.34
Table 2. PSNR and SSIM of traditional algorithm after filter.
Table 2. PSNR and SSIM of traditional algorithm after filter.
ImageAlgorithmMR = 0.25MR = 0.10MR = 0.04MR = 0.01
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
MeanTVAL316.290.4815.810.3714.730.3410.820.19
NLR-CS16.190.4615.660.3914.430.3210.610.22
D-AMP17.550.5016.860.4615.250.3210.440.17
Ours21.350.6319.920.7018.050.5015.140.34
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, H.; Lu, K.; Xue, J.; Dai, F.; Zhang, Y. Dual Optical Path Based Adaptive Compressive Sensing Imaging System. Sensors 2021, 21, 6200. https://doi.org/10.3390/s21186200

AMA Style

Li H, Lu K, Xue J, Dai F, Zhang Y. Dual Optical Path Based Adaptive Compressive Sensing Imaging System. Sensors. 2021; 21(18):6200. https://doi.org/10.3390/s21186200

Chicago/Turabian Style

Li, Hongliang, Ke Lu, Jian Xue, Feng Dai, and Yongdong Zhang. 2021. "Dual Optical Path Based Adaptive Compressive Sensing Imaging System" Sensors 21, no. 18: 6200. https://doi.org/10.3390/s21186200

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop