Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Hydrothermal Synthesis of WO3·0.33H2O Nanorod Bundles as a Highly Sensitive Cyclohexene Sensor
Next Article in Special Issue
Anti-Interference Aircraft-Tracking Method in Infrared Imagery
Previous Article in Journal
A Seabed Real-Time Sensing System for In-Situ Long-Term Multi-Parameter Observation Applications
Previous Article in Special Issue
Automatic and Robust Infrared-Visible Image Sequence Registration via Spatio-Temporal Association
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sensitivity Improvement of Extremely Low Light Scenes with RGB-NIR Multispectral Filter Array Sensor

Department of Electrical and Electronic Engineering, Yonsei University, 50 Yonsei Road, Seodaemun-gu, Seoul 03722, Korea
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(5), 1256; https://doi.org/10.3390/s19051256
Submission received: 31 January 2019 / Revised: 8 March 2019 / Accepted: 8 March 2019 / Published: 12 March 2019
(This article belongs to the Special Issue Infrared Sensors and Technologies)

Abstract

:
Recently, several red-green-blue near-infrared (RGB-NIR) multispectral filter arrays (MFAs), which include near infrared (NIR) pixels, have been proposed. For extremely low light scenes, the RGB-NIR MFA sensor has been extended to receive NIR light, by adding NIR pixels to supplement for the insufficient visible band light energy. However, the resolution reconstruction of the RGB-NIR MFA, using demosaicing and color restoration methods, is based on the correlation between the NIR pixels and the pixels of other colors; this does not improve the RGB channel sensitivity with respect to the NIR channel sensitivity. In this paper, we propose a color restored image post-processing method to improve the sensitivity and resolution of an RGB-NIR MFA. Although several linear regression based color channel reconstruction methods have taken advantage of the high sensitivity NIR channel, it is difficult to accurately estimate the linear coefficients because of the high level of noise in the color channels under extremely low light conditions. The proposed method solves this problem in three steps: guided filtering, based on the linear similarity between the NIR and color channels, edge preserving smoothing to improve the accuracy of linear coefficient estimation, and residual compensation for lost spatial resolution information. The results show that the proposed method is effective, while maintaining the NIR pixel resolution characteristics, and improving the sensitivity in terms of the signal-to-noise ratio by approximately 13 dB.

1. Introduction

Recently, several attempts have been made to utilize near infrared (NIR) band information. Multispectral images observed in various spectrum bands, including both visible and NIR bands, have been used in remote sensing applications [1,2]. As each spectral band provides a different type of information, the spectral bands were selectively used in observing the multi-spectral images. In surveillance [3] and night vision cameras [4], the NIR band is particularly useful under low lighting or invisible NIR lighting conditions. The NIR band is also used in biometric [5], face matching [6], and face recognition [7] applications, which have been studied based on the intrinsic reflectivity of the skin or eyes under NIR illumination. As the NIR reflection is material dependent, it is also used in material classification [8] and illuminant estimation [9]. NIR images can be used in image enhancement applications such as image dehazing [10].
Most digital imaging devices use a color filter array (CFA) to reduce equipment cost and size instead of using three sensors and optical beam splitters. The Bayer CFA, comprising the primary colors red, green, and blue (R, G, and B), is a widely used CFA [11]. Figure 1a shows the Bayer CFA. Recently, methods using a new CFA have been studied to overcome the limited sensitivity of the Bayer CFA under extremely low light conditions. This is because the amount of absorbed light decreases on account of the RGB color filters. Thus, sensitivity improvement methods utilizing an invisible light source have been proposed in recent years [12,13,14,15,16,17,18,19]. Because the wavelength of an invisible light source is different from that of visible light, the invisible and visible light images can be captured simultaneously. New CFA patterns [20] and demosaicing methods [21,22,23,24], containing NIR pixels in the pattern have been proposed. Figure 1b shows an example of a red-green-blue NIR (RGB-NIR) multispectral filter array (MFA). Bennett et al. [12] constructed an imaging system that can simultaneously capture color and NIR images. For a pair of color and NIR images, the proposed system outperformed the reconstruction methods reported in [25,26]. Yan et al. [16] preserved the necessary details and edges of a color image by considering the structural differences between the NIR and color images. These studies reported good results for enhancing color images of low light scenes. However, existing methods have an implied restriction, regarding how dark the environment can be. If the lighting conditions are darker than that assumed in these methods, they will not work well.
Generally, to utilize the RGB-NIR MFA, two steps are involved (Figure 2): Demosaicing and color restoration [20,27]. The demosaicing process, also called color interpolation, is a process of converting sub-sampled R, G, B, and NIR channels into full resolution images. Figure 3 shows an image taken under extremely low (0.01lx) incandescent light conditions. From Figure 3a,b, it is possible to obtain high sensitivity color images from the RGB-NIR MFA by incandescent lighting, with a large amount of NIR energy under extremely low light conditions. However, demosaiced RGB channels contain the NIR band, as shown in Figure 3a, which is different from typical RGB channel color reproduction. Therefore, a color restoration process is required to estimate and remove the included NIR band. Figure 3c shows the results of the color restoration algorithm for the image, shown in Figure 3a [27], with the brightness similar to that of the NIR channel, a gain of 10 times. The two processes, using the NIR channel from the images (Figure 3), indicate that the sensitivity of the color image is not improved with respect to the sensitivity of the NIR channel under extremely low light conditions.
Various post-processing algorithms have been studied using the NIR band for this problem. For example, image fusion methods [28] fuse the NIR and RGB channels. Fusion methods focus on identifying invisible or indistinguishable subjects in the visible band, while maintaining the RGB color components. Although object discrimination ability improved, the sensitivity remains poor. Therefore, the local contrast and color are distorted. Another technique involves using a guided filtering (GF) method [29,30] that considers the linear correlation between the luminance (Y) channel of the visible band, and the NIR channel, to reflect the sensitivity characteristics of the NIR channel. This method uses linear regression between the Y and NIR channels to estimate the linear coefficients. The NIR channel is reconstructed into Y channels by using the coefficients. This method is also used for reconstructing low illuminance images from a flashing image. However, under extremely low light conditions, a visible image with a high level of noise makes it difficult to estimate the linear coefficients accurately. Moreover, the resolution of the Y channel, reconstructed from the linear coefficients, is not up to that of the NIR channel. To solve this problem, five cost functions, and an algorithm to minimize them, have been proposed [31]. This algorithm is more stable than the above algorithms under extremely low light conditions; however, it is difficult to realize NIR channel level sensitivity because of the trade-off between the five factors in the cost minimization process.
In this paper, we propose a post-processing method that can enhance the spatial resolution, while improving the brightness and signal-to-noise ratio (SNR) of the RGB channels, by reflecting the sensitivity of the NIR channel. High sensitivity NIR channels, at very low light levels, have better spatial resolution and sensitivity than noisy RGB channels. On the other hand, RGB channels contain color information and local contrasts that represent the colors. Hence, we propose a method to improve the resolution and SNR of the R, G, and B channels by reconstructing the NIR channel. In particular, we propose a smoother method to improve the accuracy of linear coefficients of guided filtering in extremely low light conditions. In addition, we propose residual information compensation to recover spatial resolution information and texture, lost in the guided filtering process.
This paper proposes an edge preserving smoother based pre-processing algorithm to estimate the linear coefficients of the guided filter accurately. The edge preserving smoother proposed in [16] estimates the kernel from the sensitive NIR channel and applies the kernel to the Y channel. The smoothened Y channel removes the noise and texture components, that interfere with linear coefficient estimation, along the strong edges of the NIR channel. This procedure allows the guided filter to estimate the correct coefficient in extremely low light scenes. In addition, this paper proposes a post-processing algorithm to improve the resulting resolution of the guided filter. Guided filtering is used to estimate the residual component from the NIR channel and compensate for the insufficient spatial resolution of the reconstructed Y channel. The residual component is the spatial resolution information of the missing NIR channel in the guided filtering process. Finally, the proposed algorithm maintains the color and local contrast of the RGB channel, through guided filtering and outputs the result with NIR channel level sensitivity and spatial resolution.
The remainder of this paper is organized as follows. In Section 2, the research problem is detailed. In Section 3, the proposed post-processing method based on guided filtering is described. In Section 4, we present the results of conducted experiments, and compare the results of the proposed method with those of other post-processing algorithms. Our conclusions are presented in Section 5.

2. Problem Statement

The color restoration results of the RGB-NIR MFA under extremely low light conditions include a high level of spatial and color noise, as shown in Figure 3c. To analyze the characteristics of the noise, Figure 4 shows the color restoration results with respect to the type of channel. The RGB channels show a similar level of noise, whereas the noise level in the Y channel is more than that in the NIR channel. On the other hand, the NIR channel is a high sensitivity image, with low noise and accurate texture information of the subject. However, it does not have the local brightness or contrast shown in the visible band. This is because the hat and the head of the subject are not clearly distinguished.
In short, the rationale of the proposed post-processing method is that the reconstructed image has NIR channel level brightness and sensitivity. Moreover, the color image has an NIR channel level SNR and color reproduction at the color restored RGB result level. Table 1 lists the characteristics of the color restored RGB result image, obtained using the proposed post-processing algorithm. The brightness, sensitivity, noise, and spatial resolution characteristics of the image should follow those of the NIR channel. On the other hand, the hue, saturation, and local contrast must follow those of the color restored RGB channel to maintain color reproduction.
The proposed post-processing algorithm follows the guided filtering method [29]. The guided filtering estimates the linear coefficients between the NIR channel and the Y channel by using a linear regression method. From the estimated linear coefficients, the NIR channel is reconstructed into associations with local brightness and contrast similar to the Y channel. In addition, a smoothing process, such as mean operation, for the linear coefficients can help reduce the influence of noise. However, it is difficult to estimate the linear coefficients accurately because the noise under extremely low light conditions occurs, not only in the high frequency band, but also in the low frequency band. The noise generated up to the low frequency band can be removed using a strong smoothing filter. However, this will damage both the occlusions of the subject and the strong edge areas, thus reducing the accuracy of linear coefficient estimation in the strong edge regions and generating artifacts in the resulting image.
Figure 5 shows the input channels, RGB channels, NIR channel, and guided filtering results. Figure 5c shows the artifacts due to the error in estimating the linear coefficients during smoothing. In the background region, noise characteristics, amplified over the NIR channel, can be seen. This is due to the failure in accurately estimating the linear coefficients in the noise region, as shown in Figure 5a. In addition, the guided filtering result shows that there is no texture information on the hat. Therefore, the proposed algorithm performs guided filtering correction, thus establishing a post-processing algorithm suitable for extremely low light conditions.

3. Proposed Post-Processing Method Based on Guided Filtering

3.1. Post-Processing Framework

To solve the problem of noise, linear coefficients, and texture information of the NIR channel, under extremely low light conditions, this paper proposed a post-processing method based on guided filtering. The proposed algorithm comprises three steps: edge preserving smoothing, guided filtering, and residual compensation. Figure 6 shows the framework of the proposed post-processing algorithm.
First, the color restored RGB channel, converted into the YUV domain, is divided into luminance and chrominance channels, and the Y channel is inputted to the edge preserving smoother along with the NIR channel. The reason for converting to the YUV domain is to separate color components of RGB, such as saturation and hue, to prevent color change during the algorithm execution. The U and V channels, except for the luminance component, contain color components and represent color information and color noise.
The edge preserving smoother can accurately preserve the strong edges and occlusions of the object by estimating the edge from the sensitive NIR channel. The smoothing kernel, estimated from the NIR channel, has little influence on the Y channel in terms of the noise. Thus, it can remove the low frequency noise and preserve the strong edges. In addition, the color noise, existing in the U and V channels, is removed through the kernel estimated from the NIR channel, thus maintaining the correct color at the boundary of the object. In this paper, a cross bilateral filter (CBF) [26] was used as an edge preserving smoother. The NIR channel passing through the smoothing filter contains local brightness and local contrast information, and the noise and texture information can be removed. This can improve the regional distribution similarity between the Y and NIR channels in estimating the linear coefficients. Moreover, the smoothened Y and NIR channels maintain a strong edge, and the guided filter can accurately estimate the linear coefficient of change at the boundary of the object.
Guided filtering estimates the linear coefficients through linear regression. The estimated linear coefficients help reconstruct the NIR channel to create a new Y channel with improved sensitivity. The proposed algorithm smoothens the estimated linear coefficients by using the CBF kernel estimated from the NIR channel. This stabilizes the small errors that occur in the estimation process and preserves the linear coefficients in the occlusion of the object. In addition, the proposed algorithm updates the reconstructed Y channel, by estimating the texture and high frequency components (the residual information) that have not yet been compensated via guided filtering.
Finally, the Y channel, reconstructed from the NIR channel, has a local contrast of the input Y channel and local brightness characteristics. In addition, the reconstructed Y channel has NIR level sensitivity, noise, and spatial resolution characteristics. Similarly, the smoothened U and V channels are suppressed from noise in all the frequency bands, and they can maintain local color characteristics.

3.2. Edge Preserving Smoother and Linear Coefficients

The key assumption in the guided filtering is the local linear model between the NIR and Y channels. The local linear model can be expressed as follows:
Y S = a S N S + b S ,
where S is the local area of the object in an image, Y and N are the Y, and NIR channels, respectively, and a S and b S are the linear coefficients. Equation (1) can be expressed as follows for the current pixel position:
I Y ( p ) = a ( p ) I N ( p ) + b ( p ) ,                     p S ,
where I Y ( p ) and I N ( p ) are the Y and NIR channel pixel values at the p th position, respectively. The linear coefficients are assumed to be constant in S . To obtain the linear coefficients using a simple linear regression model, the following equations have been proposed [32]:
a S = C o v ( N S , Y S ) V a r ( N S ) + ε   ,
b S = M e a n ( Y S ) a S M e a n ( N S ) ,
where ε is a small constant to prevent division by zero, C o v ( N S , Y S ) is the covariance between the NIR and Y channels in the region S , V a r ( N S ) is the variance of NIR in the region S , and M e a n ( Y S ) is the mean value of the Y channel in the region S . The following equations can be used to obtain the mean, variance, and covariance from the NIR and Y channels:
M e a n S ( N ( p ) ) = p , q S I N ( q ) / | S | ,
V a r S ( N ( p ) ) = p , q S ( I N ( q ) ) 2 / | S | ( p , q S I N ( q ) / | S | ) 2 ,
C o v S ( N ( p ) , Y ( p ) ) = p , q S I N ( q )   I Y ( q ) / | S | M e a n ( N ( p ) ) · M e a n ( Y ( p ) ) ,
where q is the pixel position in the region S , | S | is the number of pixels in the region S , and M e a n ( N ( p ) ) is the mean value at the current pixel position p in the NIR channel. V a r ( N ( p ) ) is the variance value at the current pixel position p in the NIR channel, and C o v ( N ( p ) , Y ( p ) ) is the covariance value at the current pixel position p between the Y and NIR channels.
The linear coefficients of the guided filter, mentioned in Section 2, show errors because the covariance is estimated from the noisy Y channel to the NIR channel. In other words, as the noise level increases, the linearity of the Y channel with the NIR channel decreases, and the accuracy of the coefficient a S decreases. An inaccurately predicted a S value can lead to an inaccurate prediction of b S in Equation (4). In conclusion, it is necessary to improve the accuracy of C o v ( N S , Y S ) correlated with the Y channel.
A Gaussian filter is incorporated in the guided filtering method [29] to improve the accuracy of linear coefficient estimation. However, when the noise level is high, the strong edges are blurred. Therefore, this paper proposes an edge preserving smoother by pre-processing the guided filter. Figure 7 compares the various Gaussian function based smoothing filter results for the NIR and Y channels. The filter characteristics are as follows: Gaussian filter with no edge preservation, a bilateral filter with edge preserving, and a cross bilateral filter to estimate the kernel from the guide image. As described above, the Gaussian filter result (Figure 7b) shows that although the influence of the noise is reduced, the occlusion of the object is significantly blurred. In the bilateral filter results, shown in Figure 7c, the noise removal of the Y channel is limited. This is the limitation of the kernel estimated from the Y channel, which contains significant noise under extremely low light conditions, even for the edge preserving smoother. The cross bilateral filter results (Figure 7d) show that the noise is effectively removed from the Y channel, and the strong edge and occlusion regions are preserved in both the NIR and Y channels. Moreover, the texture information is effectively removed from the NIR channel. Hence, we used the CBF as a pre-processing step in the guide filtering to take advantage of the highly sensitive NIR channel characteristics; the CBF replaces the mean operation. The existing mean operation is a uniform blur method and is replaced with an edge preserving smoother. The CBF can be expressed as follows:
M e a n N , S C B F ( Y ( p ) ) = 1 W N , S C B F ( p ) p , q S G σ s ( | | p q | | ) G σ r ( | I N ( p ) I N ( q ) | ) I Y ( q ) ,
W N , S C B F ( p ) = p , q S G σ s ( | | p q | | ) G σ r ( | I N ( p ) I N ( q ) | ) ,
where M e a n N , S C B F ( Y ( p ) ) is the CBF result of the Y channel at the p th pixel, obtained by applying a bilateral filter kernel estimated from the NIR channel, G σ s and G σ r are the Gaussian functions for the spatial and range kernels, respectively, σ s and σ r are the smoothing parameters, and W N , S C B F ( p ) is the normalized factor at the p th pixel. For smoother results in Figure 7, for a 48 (16 + 16 + 16) bit color input image, the parameter settings for filters are Gaussian filter σ = 300, bilateral filter σ s = 300, bilateral filter σ r = 5000, cross bilateral filter σ s = 300, and bilateral filter σ r = 5000. The filter kernel size ( S ) is 31 × 31. The variance and covariance values estimated using the proposed CBF are as follows:
V a r N , S C B F ( N ( p ) ) = 1 W N , S , V a r C B F ( p ) p , q S ( G σ s ( | | p q | | ) G σ r ( | I N ( p ) I N ( q ) | ) I N ( q ) ) 2 M e a n N C B F ( N ( p ) ) 2 ,
W N , S , V a r C B F ( p ) = p , q S G σ s ( | | p q | | ) 2 G σ r ( | I N ( p ) I N ( q ) | ) 2 ,
C o v N , S C B F ( N ( p ) , Y ( p ) ) = 1 W N , S , V a r C B F ( p ) p , q S G σ s ( | | p q | | ) 2 G σ r ( | I N ( p ) I N ( q ) | ) 2 I N ( q ) I Y ( q ) M e a n N C B F ( N ( p ) ) · M e a n N C B F ( Y ( p ) ) ,
where V a r N , S C B F ( N ( p ) ) is the variance value of the smoothing filtered NIR channel at the p th pixel, obtained by applying a bilateral filter kernel estimated from the NIR channel, C o v N , S C B F ( N ( p ) , Y ( p ) ) is the variance value of the smoothing filtered Y channel at the p th pixel, obtained by applying a bilateral filter kernel estimated from the NIR channel, and W N , S , V a r C B F ( p ) is the normalized factor for the variance at the p th pixel. For conditions similar to the smoothed Y channels with texture information and noise removed, the NIR channel also needs to pass through the same bilateral filter kernel. This process minimizes the effect of texture on the covariance estimate and improves the accuracy of estimating the correlation between the local brightness of the objects on each channel. The linear coefficients a s C B F and b s C B F associated with the CBF can be expressed as follows:
a S C B F = C o v N , S C B F ( N ( p ) , Y ( p ) ) V a r N , S C B F ( N ( p ) ) + ε ,
b S C B F = M e a n N , S C B F ( Y ) a S M e a n N , S C B F ( N ) .
The linear coefficients estimated from the smoothed image, are more accurate than the original values, but pass through the estimated bilateral filter kernel from the NIR channel to eliminate some possible estimation errors. The linear coefficients passing through the CBF kernel are expressed as follows:
a ^ S ( p ) = M e a n N , S C B F ( a S ( p ) ) ,
b ^ S ( p ) = M e a n N , S C B F ( b S ( p ) ) ,
I Y ^ N ( p ) = a ^ ( p ) I N ( p ) + b ^ ( p ) ,
where a ^ S and b ^ S are the smoothed linear coefficients, Y ^ N is the Y channel reconstructed from the NIR channel by using the linear coefficients a ^ S and b ^ S .
Figure 8c,d show the guided filtering results, and proposed CBF + guided filtering results, respectively. The accuracy of linear coefficient estimation is improved with the use of CBF, showing stable results in the background. Figure 8d shows that the texture components on the cap are more effectively restored, compared to the guided filtering results. However, Figure 9d shows that the texture on the foot of the doll is not well restored. Figure 10d shows that the texture is not well reconstructed in the resolution chart area at the bottom of the color chart. To solve this problem, we perform post-processing of the guided filter.

3.3. Compensation with Residual Information of NIR

The texture information, in the channel reconstruction results of the guided filter, is lacking. Because of the area wise computation of the linear coefficients, it is difficult to reconstruct the texture. To obtain spatial resolution at the NIR channel level, this paper proposes a residual information compensation process. The residual information is defined as the information lost in the NIR channel among the textures resulting from guided filtering. To estimate the residual information, the NIR channel is reconstructed using the Y channel. This process traces the loss process through the channel other than the NIR channel to estimate the texture information lost during guided filtering. The process of reconstructing the NIR channel, by using the linear coefficients estimated in the previous sub section is as follows:
I N ^ Y ( p ) = I Y ( p ) b ^ ( p ) a ^ ( p ) + ε ,
where N ^ Y is the NIR channel reconstructed from the Y channel. Equation (18) is a transformed version of Equation (17). The residual information is estimated as follows:
R N N ^ Y = N N ^ Y ,
where R N N ^ Y is the residual information and the difference between the NIR channel and the channel N ^ Y reconstructed from the Y channel. The process of compensating for the residual information is as follows:
Y o u t = N ^ N + R N N ^ Y ,
where Y o u t is the proposed result of the luminance channel. The results of the chrominance channels are expressed as:
U o u t = M e a n N , S C B F ( U ) ,
V o u t = M e a n N , S C B F ( V ) ,
where U o u t and V o u t are the resulting chrominance channels of the proposed algorithm, and U and V are the chrominance channels of the input image.
Figure 8e, Figure 9e, and Figure 10e show the GF + CBF results in compensating for the residual (residual compensation, RC) information. Figure 8e shows that the white line artifact, shown in Figure 8d, is suppressed, and the texture becomes richer. Figure 9e and Figure 10e show that the missing texture information is well compensated from Figure 9d and Figure 10d.

4. Experimental Results

To verify the performance of the proposed algorithm under extremely low light conditions, we used directly captured images. The image, shown in Figure 3a, was taken in 0.01 lx incandescent light under extremely low light conditions. We measured the R, G, B, and NIR channels with full resolution by using a multi-spectral filter wheel with a single sensor. The resolution of the image used in the experiment was 1600 × 1200, and it was photographed with 16 bits per pixel of one channel. By using the color restoration algorithm proposed in [27], we obtained the RGB image, shown in Figure 3c, with a gain of 10 times gain. The proposed post-processing uses the RGB color image (Figure 3c) and the NIR gray image (Figure 3b) as the input image.
In this section, we compare various conventional methods (CM) to verify the performance of the proposed algorithm. First, we compared the core process guided filtering (CM1) [29]. This algorithm is widely used as a sensitivity enhancement algorithm that utilizes the NIR channel under general illumination conditions rather than under extremely low light conditions. Second, it was compared with the image fusion method (CM2) [28]. This method helps improve the information amount by fusing the visible and invisible bands to facilitate the material classification using the NIR channel. Finally, it is compared with the cost minimization method (CM3) [31]. This method proposes five cost functions that compensate the sensitivity characteristics of the NIR channel while maintaining the color characteristics of the RGB channel under extremely low light conditions and obtain the results through a minimization process. The experimental environment used to measure the computing time included C++ code on Intel i7-6700k CPU 4 Ghz, Windows 10, and Visual Studio 2015. The algorithm execution times for CM1, CM2, CM3, and the proposed method are 0.880 s, 1.213 s, 4.520 s, and 1.107 s, respectively. The proposed method is faster than the other algorithms except CM1, which has fast computation time because it has guided filtering included in the proposed method.
Figure 11 shows the input image, results of the three conventional methods, and the full image obtained using the proposed algorithm. In Section 2, the results of CM1 obtained using guided filtering, were poor because the linear coefficients could not be accurately determined under extremely low light conditions. The results of CM2 show that the noise characteristics of the NIR are generally followed. However, based on the local contrast and brightness of the NIR, the color reproduction of the RGB channel can’t be followed. The result of CM3 follows the color reproduction of RGB; however, the low frequency noise that occurs under extremely low light conditions is not removed, and the noise characteristic of the NIR level is not exhibited. The proposed method (PM) results follow the RGB color reproduction and show NIR level noise level at the same time. Thus, the proposed algorithm shows better noise suppression, and color reproduction ability, than conventional methods.
Figure 12 shows the image of the color chart and text area shown in Figure 11. First, the proposed algorithm for the writing area shows the expressive power of the NIR level, and no noise is amplified in the background. This was achieved by compensating for the residual information. In the color patch, only CM2 and PM are effectively expressed without the influence of noise. However, as CM2 follows the local brightness of the NIR, the color reproduction power falls short of the PM result. The good noise characteristics and color reproduction ability of the PM can be attributed to the proposed edge preserving smoother.
Figure 13 shows the image of the doll’s body area shown in Figure 11. The experimental results show that CM3 and PM effectively portray the RGB colors in terms of the color reproduction ability. However, the texture information regarding the doll’s body, and foot reflected by CM2 and PM, are not up to the mark. As shown in Figure 12, the edge preserving smoothing of the proposed algorithm helps improve the accuracy of estimating the linear coefficients via guided filtering, and the compensated residual information helps improve the texture information.
SNR [33] was used for the numerical comparison of sensitivity improvement in extremely low light conditions. The equation of SNR is expressed as:
S N R d B = 20 l o g 10 μ I σ I   ,
where S N R d B represents the SNR, and the unit is decibels (dB) in the log scale. In addition, μ I is the average of the measurement region in the image, and σ I denotes the standard deviation of the measurement region in the image.
Figure 14 shows the white flat area and the SNR were measured in this area. The SNR values of the color images in Figure 14 represent average values measured on the R, G, and B channels, respectively. The NIR channel is a single SNR value. Figure 14 shows the best noise characteristics of PM and CM2. The fusion method results reflect the local brightness of the NIR, which improves the noise characteristics. The proposed algorithm follows the local brightness of RGB and conforms more closely to the development purpose of following the noise characteristic of NIR. Table 2 shows the results of Figure 14 in more detail. The average SNR of the input image is 13.76 dB, which means that the noise is considerably large. The CM1 result has a higher SNR than RGB but the lowest value among other comparison algorithms. CM2 has a 12.97 dB improvement over RGB and higher SNR than NIR. This is a result of faithfully reflecting the NIR information. The result of CM3 is lower than that of fusion but it is 12.66 dB better than that of RGB. It faithfully follows the local brightness of RGB, but it does not suppress low frequency noise. The proposed method shows the highest SNR and 13.03 dB improvement over RGB. This is higher than that of the fusion method, but the difference is insignificant. The high numerical value of the proposed algorithm shows that the sensitivity characteristic of NIR is reflected in the fusion level.
In addition to incandescent lamps, the same experiment was conducted in an extremely low illumination environment, including the NIR band such as a sodium lamp. This is to verify the proposed method in different lighting conditions (color temperature and lighting spectrum). Fluorescent lamps or LED lamps are excluded from the comparison because they cannot utilize the NIR band without artificially adding NIR bands. Figure 15 shows the result of improving sensitivity using the NIR channel with a sodium lamp at 0.01 lx. Figure 15a shows the result of the color restoration algorithm as an input image. Figure 15b shows that the representation of the black region is awkward, due to inaccurate linear coefficient estimates, and white line artifacts occur at strong edges. In Figure 15c, the effect of noise is greatly reduced by NIR, but a color reproduction error occurs. Figure 15d shows that color reproduction is accurate and there are no artifacts in guided filtering, but it is not enough to follow the noise characteristics of NIR. Finally, Figure 15e shows that the proposed algorithm has excellent color reproduction, spatial resolution, and noise characteristics. This is similar to the comparison of results of incandescent lamps.

5. Conclusions

We used an RGB-NIR MFA to replace the existing Bayer CFA. However, this method has limitations in reflecting the sensitivity of the NIR channel through the demosaic and color restoration algorithms. To solve this problem, post-processing is proposed, and an algorithm is developed to reflect three characteristics: SNR and color noise suppression at the NIR channel level, improved resolution at the NIR channel level, and RGB color image level color reproduction. For the three characteristics, we propose an edge preserving smoother that is most suitable under extremely low light conditions, by pre-processing a guided filter and estimate the linear coefficient more accurately. The edge preserving smoother helps detect the quality degradation component of the RGB-NIR MFA, such as the noise in the RGB channels. Moreover, a CBF kernel estimated from the NIR channel is applied to the U and V channels to remove color noise. The residual information is compensated by post-processing the guided filter. The proposed algorithm can estimate and compensate for the missing texture information of the guided filter result and the information of the erroneously reconstructed strong edge regions from the NIR channel. The proposed method can estimate the residual information lost in the guided filtering process by reconstructing the luminance channel into an NIR channel by using a linear coefficient. The compensation of the residual information helps improve the resolution of the output image and correct the artifacts in the strong edge areas. The experimental results show that the proposed algorithm results in a resolution and SNR similar to that of the NIR channel, under extremely low light conditions, and strictly follows the colors of the RGB channels.

Author Contributions

S.J. designed and performed the experiments, analyzed the data, and wrote the manuscript, under the supervision of M.G.K.

Acknowledgments

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2019R1A2C2002167).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pohl, C.; Genderen, J.L.V. Review article Multisensor image fusion in remote sensing: Concepts, methods and applications. Int. J. Remote Sens. 1998, 19, 823–854. [Google Scholar] [CrossRef]
  2. Choi, J.; Yu, K.; Kim, Y. A New Adaptive Component-Substitution-Based Satellite Image Fusion by Using Partial Replacement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 295–309. [Google Scholar] [CrossRef]
  3. Hao, X.; Chen, H.; Yao, C.; Yang, N.; Bi, H.; Wang, C. A near-infrared imaging method for capturing the interior of a vehicle through windshield. In Proceedings of the 2010 IEEE Southwest Symposium on Image Analysis & Interpretation (SSIAI), Austin, TX, USA, 23–25 May 2010. [Google Scholar]
  4. Hertel, D.; Marechal, H.; Tefera, D.A.; Fan, W.; Hicks, R. A low-cost VIS-NIR true color night vision video system based on a wide dynamic range CMOS imager. In Proceedings of the 2009 IEEE Intelligent Vehicles Symposium, Xi’an, China, 3–5 June 2009; pp. 273–278. [Google Scholar]
  5. Kumar, A.; Prathyusha, K. Personal Authentication Using Hand Vein Triangulation and Knuckle Shape. IEEE Trans. Image Process. 2009, 18, 2127–2136. [Google Scholar] [CrossRef] [PubMed]
  6. Yi, D.; Liu, R.; Chu, R.; Lei, Z.; Li, S.Z. Face Matching Between Near Infrared and Visible Light Images. In Proceedings of the International Conference on Biometrics, Seoul, Korea, 27–29 August 2007; pp. 523–530. [Google Scholar]
  7. Li, S.Z.; Chu, R.; Liao, S.; Zhang, L. Illumination invariant face recognition using near-infrared images. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 627–639. [Google Scholar] [CrossRef] [PubMed]
  8. Salamati, N.; Fredembach, C.; Süsstrunk, S. Material classification using color and NIR images. In Color and Imaging Conference; Society for Imaging Science and Technology: Springfield, VA, USA, 2009; pp. 216–222. [Google Scholar]
  9. Fredembach, C.; Susstrunk, S. Illuminant estimation and detection using near-infrared. In Digital Photography V; International Society for Optics and Photonics: Bellingham, WA, USA, 2009. [Google Scholar]
  10. Schaul, L.; Fredembach, C.; Susstrunk, S. Color image dehazing using the near-infrared. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009. [Google Scholar]
  11. Bayer, B.E. Color Imaging Array. U.S. Patent No. 3,971,065, 20 July 1976. [Google Scholar]
  12. Bennett, E.P.; Mason, J.L.; Mcmillan, L. Multispectral Bilateral Video Fusion. IEEE Trans. Image Process. 2007, 16, 1185–1194. [Google Scholar] [CrossRef] [PubMed]
  13. Feng, C.; Zhuo, S.; Zhang, X.; Shen, L.; Susstrunk, S. Near-infrared guided color image dehazing. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013. [Google Scholar]
  14. Krishnan, D.; Fergus, R. Dark flash photography. ACM Trans. Graph. 2009, 28, 96–101. [Google Scholar] [CrossRef]
  15. Wu, C.; Samadani, R.; Gunawardane, P. Same frame rate IR to enhance visible video conference lighting. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011. [Google Scholar]
  16. Yan, Q.; Shen, X.; Xu, L.; Zhuo, S.; Zhang, X.; Shen, L.; Jia, J. Cross-Field Joint Image Restoration via Scale Map. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Sydney, Australia, 1–8 December 2013. [Google Scholar]
  17. Zhang, X.; Sim, T.; Miao, X. Enhancing photographs with Near Infra-Red images. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008. [Google Scholar]
  18. Zhuo, S.; Zhang, X.; Miao, X.; Sim, T. Enhancing low light images using near infrared flash images. In Proceedings of the 2010 IEEE International Conference on Image, Hong Kong, China, 26–29 September 2010. [Google Scholar]
  19. Matsui, S.; Okabe, T.; Shimano, M.; Sato, Y. Image Enhancement of Low-light Scenes with Near-infrared Flash Images. IPSJ Trans. Comput. Vis. Appl. 2010, 2, 215–223. [Google Scholar] [CrossRef]
  20. Tian, Q.; Lansel, S.; Farrell, J.E.; Wandell, B.A. Automating the design of image processing pipelines for novel color filter arrays: Local, linear, learned (L3) method. In Digital Photography X; International Society for Optics and Photonics: Bellingham, WA, USA, 2014. [Google Scholar]
  21. Condat, L. A generic variational approach for demosaicking from an arbitrary color filter array. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009. [Google Scholar]
  22. Gu, J.; Wolfe, P.J.; Hirakawa, K. Filterbank-based universal demosaicking. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010. [Google Scholar]
  23. Park, S.W.; Kang, M.G. Generalized color interpolation scheme based on intermediate quincuncial pattern. J. Electron. Imaging 2014, 23, 030501. [Google Scholar] [CrossRef]
  24. Petrovic, V.; Xydeas, C. Gradient-Based Multiresolution Image Fusion. IEEE Trans. Image Process. 2004, 13, 228–237. [Google Scholar] [CrossRef] [PubMed]
  25. Eisemann, E.; Durand, F. Flash photography enhancement via intrinsic relighting. ACM Trans. Graph. 2004, 23, 673–678. [Google Scholar] [CrossRef]
  26. Petschnigg, G.; Szeliski, R.; Agrawala, M.; Cohen, M.; Hoppe, H.; Toyama, K. Digital photography with flash and no-flash image pairs. ACM Trans. Graph. 2004, 23, 664–672. [Google Scholar] [CrossRef]
  27. Park, C.; Kang, M. Color Restoration of RGBN Multispectral Filter Array Sensor Images Based on Spectral Decomposition. Sensors 2016, 16, 719. [Google Scholar] [CrossRef] [PubMed]
  28. Li, S.; Kang, X.; Fang, L.; Hu, J.; Yin, H. Pixel-level image fusion: A survey of the state of the art. Inf. Fusion 2017, 33, 100–112. [Google Scholar] [CrossRef]
  29. He, K.; Sun, J.; Tang, X. Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 6, 1397–1409. [Google Scholar] [CrossRef] [PubMed]
  30. Monno, Y.; Teranaka, H.; Yoshizaki, K.; Tanaka, M.; Okutomi, M. Single-Sensor RGB-NIR Imaging: High-Quality System Design and Prototype Implementation. IEEE Sens. J. 2019, 19, 497–507. [Google Scholar] [CrossRef]
  31. Sugimura, D.; Mikami, T.; Yamashita, H.; Hamamoto, T. Enhancing Color Images of Extremely Low Light Scenes Based on RGB/NIR Images Acquisition With Different Exposure Times. IEEE Trans. Image Process. 2015, 24, 3586–3597. [Google Scholar] [CrossRef] [PubMed]
  32. Montgomery, D.C.; Peck, E.A.; Vining, G.G. Simple linear regression. In Introduction to Linear Regression Analysis, 5th ed.; John Wiley & Sons: Hoboken, NJ, USA, 2012; ISBN 978-0-470-54281-1. [Google Scholar]
  33. Devi, M.S.; Sukumar, R. Metaheuristic Based Noise Identification and Image Denoising Using Adaptive Block Selection Based Filtering. Circuits Syst. 2016, 7, 2729–2751. [Google Scholar] [CrossRef]
Figure 1. (a) Bayer color filter array [11]; (b) RGB-NIR multispectral filter array [27].
Figure 1. (a) Bayer color filter array [11]; (b) RGB-NIR multispectral filter array [27].
Sensors 19 01256 g001
Figure 2. RGB-NIR multispectral filter array system framework.
Figure 2. RGB-NIR multispectral filter array system framework.
Sensors 19 01256 g002
Figure 3. Color restored results [27] under extremely low light condition (0.01lx, incandescent); (a) input RGB image; (b) input NIR image; (c) 10 times gained image from color restored RGB.
Figure 3. Color restored results [27] under extremely low light condition (0.01lx, incandescent); (a) input RGB image; (b) input NIR image; (c) 10 times gained image from color restored RGB.
Sensors 19 01256 g003
Figure 4. Color restoration results by channel [27] under extremely low light condition (0.01lx, incandescent).
Figure 4. Color restoration results by channel [27] under extremely low light condition (0.01lx, incandescent).
Sensors 19 01256 g004
Figure 5. Guided filtering [29] results under extremely low light condition (0.01lx, incandescent); (a) input RGB image; (b) input NIR image; (c) the post-processing result by guided filtering (GF).
Figure 5. Guided filtering [29] results under extremely low light condition (0.01lx, incandescent); (a) input RGB image; (b) input NIR image; (c) the post-processing result by guided filtering (GF).
Sensors 19 01256 g005
Figure 6. The framework of proposed post-processing.
Figure 6. The framework of proposed post-processing.
Sensors 19 01256 g006
Figure 7. The comparison of various smoother results; (a) input channels; (b) Gaussian filtering results; (c) bilateral filtering results; (d) cross bilateral filtering results.
Figure 7. The comparison of various smoother results; (a) input channels; (b) Gaussian filtering results; (c) bilateral filtering results; (d) cross bilateral filtering results.
Sensors 19 01256 g007
Figure 8. Comparison of step-by-step results of the proposed post-processing; (a) input RGB; (b) input NIR; (c) guided filter (GF) result; (d) the result of using cross bilateral filter (CBF) as the post processing of guided filtering; (e) compensation of residual (RC) information to (d).
Figure 8. Comparison of step-by-step results of the proposed post-processing; (a) input RGB; (b) input NIR; (c) guided filter (GF) result; (d) the result of using cross bilateral filter (CBF) as the post processing of guided filtering; (e) compensation of residual (RC) information to (d).
Sensors 19 01256 g008
Figure 9. Comparison of step-by-step results of the proposed post-processing; (a) input RGB; (b) input NIR; (c) guided filter (GF) result; (d) the result of using cross bilateral filter (CBF) as the post processing of guided filtering; (e) compensation of residual (RC) information to (d).
Figure 9. Comparison of step-by-step results of the proposed post-processing; (a) input RGB; (b) input NIR; (c) guided filter (GF) result; (d) the result of using cross bilateral filter (CBF) as the post processing of guided filtering; (e) compensation of residual (RC) information to (d).
Sensors 19 01256 g009
Figure 10. Comparison of step-by-step results of the proposed post-processing; (a) input RGB; (b) input NIR; (c) guided filter (GF) result; (d) the result of using cross bilateral filter (CBF) as the post processing of guided filtering; (e) compensation of residual (RC) information to (d).
Figure 10. Comparison of step-by-step results of the proposed post-processing; (a) input RGB; (b) input NIR; (c) guided filter (GF) result; (d) the result of using cross bilateral filter (CBF) as the post processing of guided filtering; (e) compensation of residual (RC) information to (d).
Sensors 19 01256 g010
Figure 11. Comparison of various post- processing results (whole images); (a) input RGB; (b) input NIR; (c) conventional method 1 [29]; (d) conventional method 2 [28]; (e) conventional method 3 [31]; (f) proposed method.
Figure 11. Comparison of various post- processing results (whole images); (a) input RGB; (b) input NIR; (c) conventional method 1 [29]; (d) conventional method 2 [28]; (e) conventional method 3 [31]; (f) proposed method.
Sensors 19 01256 g011
Figure 12. Comparison of various post-processing results (cropped images); (a) input RGB; (b) input NIR; (c) conventional method 1 [29]; (d) conventional method 2 [28]; (e) conventional method 3 [31]; (f) proposed method.
Figure 12. Comparison of various post-processing results (cropped images); (a) input RGB; (b) input NIR; (c) conventional method 1 [29]; (d) conventional method 2 [28]; (e) conventional method 3 [31]; (f) proposed method.
Sensors 19 01256 g012
Figure 13. Comparison of various post- processing results (cropped images); (a) input RGB; (b) input NIR; (c) conventional method 1 [29]; (d) conventional method 2 [28]; (e) conventional method 3 [31]; (f) proposed method.
Figure 13. Comparison of various post- processing results (cropped images); (a) input RGB; (b) input NIR; (c) conventional method 1 [29]; (d) conventional method 2 [28]; (e) conventional method 3 [31]; (f) proposed method.
Sensors 19 01256 g013
Figure 14. Comparison of various post- processing results (cropped images); (a) input RGB; (b) input NIR; (c) conventional method 1 [29]; (d) conventional method 2 [28]; (e) conventional method 3 [31]; (f) proposed method.
Figure 14. Comparison of various post- processing results (cropped images); (a) input RGB; (b) input NIR; (c) conventional method 1 [29]; (d) conventional method 2 [28]; (e) conventional method 3 [31]; (f) proposed method.
Sensors 19 01256 g014
Figure 15. Comparison of various post- processing results (sodium lamp, 0.01lx, cropped images); (a) input RGB; (b) conventional method 1 [29]; (c) conventional method 2 [28]; (d) conventional method 3 [31]; (e) proposed method.
Figure 15. Comparison of various post- processing results (sodium lamp, 0.01lx, cropped images); (a) input RGB; (b) conventional method 1 [29]; (c) conventional method 2 [28]; (d) conventional method 3 [31]; (e) proposed method.
Sensors 19 01256 g015
Table 1. Characteristics of red-green-blue (RGB) and near infrared (NIR) channels. (o means include, x means not include).
Table 1. Characteristics of red-green-blue (RGB) and near infrared (NIR) channels. (o means include, x means not include).
CharacteristicsRGBNIR
Sensitivity and brightnessxo
Noisexo
Resolutionxo
Local contrast for colorox
Hue & saturationox
Table 2. Comparisons of signal-to-noise ratio (SNR) under extremely low light condition (0.01 lx).
Table 2. Comparisons of signal-to-noise ratio (SNR) under extremely low light condition (0.01 lx).
0.01lxInput RGBInput NIRCM1 [29]CM2 [28]CM3 [30]PM
R14.5826.4625.9226.4126.4127.21
G13.00-25.9127.0026.4226.45
B13.69-25.9826.7926.4326.71
Ave.13.7626.4625.9426.7326.4226.79

Share and Cite

MDPI and ACS Style

Jee, S.; Kang, M.G. Sensitivity Improvement of Extremely Low Light Scenes with RGB-NIR Multispectral Filter Array Sensor. Sensors 2019, 19, 1256. https://doi.org/10.3390/s19051256

AMA Style

Jee S, Kang MG. Sensitivity Improvement of Extremely Low Light Scenes with RGB-NIR Multispectral Filter Array Sensor. Sensors. 2019; 19(5):1256. https://doi.org/10.3390/s19051256

Chicago/Turabian Style

Jee, Seunghoon, and Moon Gi Kang. 2019. "Sensitivity Improvement of Extremely Low Light Scenes with RGB-NIR Multispectral Filter Array Sensor" Sensors 19, no. 5: 1256. https://doi.org/10.3390/s19051256

APA Style

Jee, S., & Kang, M. G. (2019). Sensitivity Improvement of Extremely Low Light Scenes with RGB-NIR Multispectral Filter Array Sensor. Sensors, 19(5), 1256. https://doi.org/10.3390/s19051256

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop