|
1.IntroductionTo reduce cost, when a digital camera captures a color image, only one color channel out of three color channels is captured. The single charge-coupled device sensor estimates the amount of red, green, and blue in the pixel. This sensor is covered with a color filter array (CFA), which causes each pixel in the camera to capture only one color channel. As pixels only have single color information, the other two color components are reconstructed, and this process is called demosaicking. One of the most famous CFA patterns called Bayer CFA is shown in Fig. 1. In this pattern, each group has nine pixels, in which five are green pixels located in a quincunx manner, and the other four pixels are red and green pixels. Based on the Bayer pattern, various demosaicking methods were proposed in the past few decades.1,2 These methods were developed to acquire a full-color image by utilizing the color differences among three color planes in the spatial or frequency domains. Especially, the adaptive color plane interpolation is an early technique that uses an edge sensing interpolation.3 This method uses three predictors and then chooses one of them via an edge classifier using the Laplacian second-order derivatives for the color components and the gradient term for the green component. Based on this approach, several methods were proposed. The adaptive filtering for CFA demosaicking (AFD) uses frequency analysis of Bayer CFA samples and tries to extract high-frequency information by a properly designed filter.4 Based on this analysis, AFD applies a low-pass filter for the luminance of green positions to better reconstruct high-frequency information. The high-frequency information at the red/blue positions is then reconstructed by adaptive filtering of color difference components. Color interpolation using the variance of color differences (VCD) is a heuristic approach used to determine a distinct edge;5 in this approach, the missing green values are grouped into plane and edge regions. This method uses the heuristic demosaicking approach. The high-order interpolation (HOI) method is based on a Taylor series that uses a weighted median filter to select a predictor.6 This method consists of two steps: (1) interpolation to determine four estimates of a missing color and (2) an edge orientation map-based classifier. In Ref. 7, the authors proposed a edge strength filter that provides local, orientation-free luminance transition information. The edge strength filter is useful in recognizing the areas where a constant color difference assumption may fail, thus improving the demosaicking performance. By adding fixed weights to Ref. 8, an effective demosaicking method based on the edge property (EDAEP) was proposed.9 The authors used an accurate edge detecting method to general different weights to minimize false colors in the edge areas. In Ref. 10, authors proposed to combine voting-based directional interpolation (VDI) with weighting-based interpolation for a more effective interpolation without estimating the interpolation directions erroneously in a complex region. The VDI is based on a gradient-inverse weighted interpolation along the interpolation direction as determined by a voting approach. The effective demosaicking based on subband correlation (EDUSC) was proposed using a discrete wavelet transform to discern edge pixels.11 The EDUSC method uses a subimage subband correlation to enable a good initial interpolation and accurate edge detection. The EDUSC method has three parts: (1) initial interpolation in each color channel, (2) missing pixel calculation from two other color channels, and (3) iterative subband synthesis for artifact reduction. In minimized-Laplacian residual interpolation (MLRI),12 the authors estimated the tentative pixel values by minimizing the Laplacian energies of the residuals. Then, the authors incorporated the MLRI into the gradient-based threshold-free algorithm to restore the demosaicking results more precisely. In this study, we propose a Bayesian approach-based demosaicking algorithm (BD) with a hybrid method in consideration of pointwise and pathwise similarity. The demosaicking algorithm recovers the underlying original true color image from a mosaicked image. We formulate this method as a maximum a posteriori (MAP) problem. Some statistical justification should be investigated to solve the MAP problem. Therefore, we present a statistical interpretation and a Bayesian motivation to clarify the MAP issue. The MAP problem focuses on the distribution of the residual value between the missing color pixel and its neighboring color pixels. We examine the pointwise and patchwise models: the pointwise model preserves a large-scale structure, and the patchwise model provides a geometric outline and a fine-scale structure in a large search region. In the computational complexity concept, the pointwise model is simpler and more efficient than the patchwise model. To utilize the different advantages of point and patch structures, we design an adaptive hybrid model as a linear combination of the two models, and the parameter is adaptive to a standard deviation representing the structural feature. After the interpolation of the missing green channel, we estimate the red/blue channel by utilizing the uniform color difference in a local region. Afterward, we apply demosaicking algorithms to the Gaussian and Poisson noisy images to prove the robustness of the proposed algorithm. The rest of the paper is organized as follows: in Sec. 2, we describe the proposed algorithm. In Sec. 3, some experimental results and the corresponding discussion are provided. Finally, the conclusion is drawn in Sec. 4. 2.Proposed Algorithm2.1.Bayesian Approach-Based AnalysisLet a mosaicked image be , where is the image coordinate index. We assume that the true original color image is highly related to , as shown in Fig. 2. An estimator is considered a Bayes estimator if it minimizes among all estimators. It is equivalent to an MAP solution, in which is a loss function used as some function of the difference between estimated and true values. The quadratic loss function is commonly used in signal processing, and is a constant. The optimal Bayesian estimator (OBE) is determined by minimizing as Based on the assumption of the quadratic loss function,13 the OBE is where is the large space of all configurations of ( if ). Here parameter is the size of the data. By using the Bayes rules and the law of total probability, we haveThen, we can modify Eq. (2) into As shown in Eq. (3), to obtain the optimal estimation , we determine the distribution of and the distribution of known as a prior distribution. To derive the distributions, we need to utilize the observation in a local neighbor region centered at the to-be-interpolated pixel instead of the pixels at the same position. In the local region , we assume that the prior distribution of is a uniform distribution, that is, . In this local neighbor region, we assume that the observation has similar structures. To achieve the optimal estimation, we use the following approximation:13 Then, Eq. (3) is changed into where can be any neighbor pixels in a square region . In this paper, we choose for high efficiency. The target is changed to find the distribution of . Figures 3(b)–3(d) show that it does not have the obvious statistical property for all the three color channels. Based on the histogram of residual images, we derive the distribution of instead of . Figures 3(e)–3(g) indicate that has a clear statistical feature. Two well-known distributions, namely, Gaussian and Laplacian, have this kind of form. We compare these two kinds of distributions with different parameters in Fig. 4. The Laplacian distribution fits the statistical property better than the Gaussian distribution. The distribution of follows the Laplacian distribution.Equation (5) is modified into where .The weight parameter plays an important role in the statistic model. For a different weight parameter , the matching has varying features, as shown on the left in Fig. 4. In the figure, we depict the empirical histogram of residual images and compare the empirical histogram with the prior statistical model of varying parameters. From the top-left to the bottom-left, the weight parameter is 0.35, 0.36, and 0.4. The top one represents the parameters with the smallest , which achieves good matching at zero but fails at tails. The bottom one represents the parameters with the largest , which achieves good matching at tails but fails at zero. In Fig. 4, the width plays a more important role than height because when the residual is zero, this pixel is in the smooth region, which is easy to be reconstructed with conventional methods. In terms of image restoration, the focus is on the complex region such as edges or textures, that is, the tails of the distribution. The tails are controlled by the width. To achieve the best matching with the empirical histogram, we should assign adaptive weight parameters to different pixel values. That is, large residual values should have large weight parameters, and small residual values should have small ones. Therefore, we design an adaptive weight parameter that is proportional to the variance of the to-be-interpolated pixel in a local region where is the expected value in the local region with size . Through Eq. (6), we can determine the original image to obtain the optimal estimation of the color image. The original image is partially unknown, but it is also partially given in the mosaicked image. Thus, in Eq. (6), we substitute the unknown into the estimated direction value of the given mosaicked color pixel values.2.2.Green Channel Reconstruction Using a Bayesian FrameworkAccording to the discussion in Sec. 2.1, we first restore the green channel and then reconstruct the red/blue channel based on the entire green channel and the given mosaicking color pixels. Figure 1 shows that a missing green color pixel at a given red/blue color pixel is surrounded by the four nearest given green color pixels in the north, south, west, and east directions. We still have the red/blue color pixel in the missing green color pixel. Therefore, we should consider this feature to reconstruct the missing green color pixel. The mosaicked image has a special feature that has pixel values in all locations. Based on this specific property of mosaicking image, instead of directly using the green color pixels in the four directions in Eq. (6), we adopt the Taylor approximation14 to provide the approximated four-direction color values by the given color pixel values In Eq. (6), we have two candidates for distance , namely, pointwise and patchwise. The patchwise distance aims to preserve the structure of the images in a large local region. In a small search region, especially in terms of image restoration, pointwise similarity plays an important role in addressing a large-scale structure. Both pointwise and patchwise distances should be considered to estimate the missing color pixels. Therefore, we propose the following hybrid Bayesian estimator (HBE)-BD algorithm: where and are the patch vectors containing the given center red pixel and four neighboring green pixels. The neighboring region is shown in Fig. 5. In this algorithm, .For images with complex features, such as various directions or textures, Eq. (6) can obtain good performance. For the two special cases of horizontal and vertical edges, we should apply the given pixels in the horizontal or vertical direction to reconstruct the missing color pixel. Horizontal and vertical directions exist in most images and occupy a large proportion in one image. Therefore, we estimate the two directions by modifying Eq. (6) and adopting the horizontal and vertical components as follows: where and are the regions containing the given color pixels in the horizontal and vertical directions and with size and (Fig. 5), respectively. Here, is adopted in Eq. (9). Then, a criterion is designed to distinguish the direction as follows: where is the threshold, and and are normalized. The interpolated full green channel is then obtained.2.3.Red/Blue Plane ReconstructionAs we fully obtained a green plane in Sec. 2.2, we now populate red/blue pixels in blue/red or green locations. In the reconstruction of the red/blue color channel, we use the restored green color channel because of the strong correlation among the color channels. To reconstruct the missing red pixels in the given blue positions, four given red pixels are located around the given blue pixel (Fig. 1). We use the color difference based on the reconstructed green pixels as follows: where and are the patch vectors including the four given red pixels located in the northwest, northeast, southwest, and southeast, and contains the same four given red pixels with size as shown in Fig. 5(c).For the missing red/blue pixel in the given pixel, Eq. (13) is applied as shown in Fig. 1. where contains the two given red pixels above and below the given green pixel.2.4.Refinement of Each Color ChannelOnce the green channel is populated using Eq. (11), for the given red pixels, the green channel can be further refined as where is the weight parameter assumed to be 0.1. Then, we refine the red/blue channel by the modified green channel as in Sec. 2.2. We iteratively refine the green channel and red/blue channel by using the above procedure until we obtain the optimal results or reach the stop condition.3.Experimental ResultsThis section presents the detailed performance comparison. Simulations were conducted on size Kodak dataset15 and two high-frequency images16 to assess the superiority of the proposed method compared with the conventional methods. The test images are shown in Fig. 6. Eight methods were selected as benchmarks: AFD,4 VCD,5 HOI,6 ESF,7 EDAEP,8 VDI,10 EDUSC,11 and MLRI.12 We evaluated the objective performance in terms of CPSNR, S-CIELAB , FSIM, and the zipper effect (ZE) aside from the visual performance. Afterward, we apply the Gaussian and Poisson noise to the Kodak dataset and provide the CPSNR results through a comparison with the existing demosaicking methods. Our simulation was conducted on an Intel Core 2 Duo CPU E8500 @ 3.16 GHz. The window size was set to 2. 3.1.Objective Performance AnalysisWe first compare the objective performances measured in terms of CPSNR. Table 1 displays the comparison of the CPSNR results of the presented method with those of other benchmark methods. Seven methods were selected as benchmarks: AFD, VCD, HOI, ESF, EDAEP, VDI, and MLRI. As shown in Table 1, the proposed method has the best average CPSNR, and the gain are 1.229 (AFD), 0.916 (VCD), 1.473 (HOI), 0.519 (ESF), 2.786 (EDAEP), 2.699 (VDI), and 1.233 (MLRI). Although images 4, 15, 16, 23, and 24 did not show the best CPSNR using the proposed method, our method still ranks second. “Diff.” shows the CPSNR difference between BD and the corresponding method. “Rank” indicates the rank of the BD method compared with those of the benchmark methods. The best performances are marked in bold. Table 1CPSNR performance comparison on the Kodak images (the best performances are marked in bold).
To numerically evaluate the results of visual performance, we used the S-CIELAB metric.17 S-CIELAB metric is a spatial form of the CIELAB developed for determining the distance between the S-CIELAB representation of an original and that of the reconstructed image. In terms of the S-CIELAB measure in the Kodak dataset, the BD method provides higher performance on average than all the other methods (AFD, VCD, HOI, ESF, EDAEP, VDI, and MLRI) with factors of 0.0522, 0.0450, 0.0703, 0.0337, 0.1789, 0.1319, and 0.0612, respectively, as shown in Table 2. Table 2S-CIELAB ΔE* results on Kodak dataset for several demosaicking methods (the best performances are marked in bold).
To further evaluate the performance of the BD method, we adopted a third metric called the FSIM index.18 The best FSIM result is 1.0, whereas a smaller FSIM indicates poor visual quality. Table 3 presents the numerical FSIM results on the Kodak dataset for various benchmarks. The BD method outperforms the other methods (AFD, VCD, HOI, ESF, EDAEP, VDI, and MLRI) at 0.0003, 0.0002, 0.0006, 0.0001, 0.0009, 0.0010, and 0.0003 on average, respectively. Table 3FSIM results on the Kodak dataset for several demosaicking methods (the best performances are marked in bold).
The ZE is one of the annoying artifacts in demosaicking. Lu and Tan introduced a measurement to evaluate the existence of the ZE. The result of the ZE represents the percentage of the pixel containing the ZE. Thus, the smaller the number is, the fewer the images containing the ZE. Table 4 shows that the BD method generated the smallest ZE. The BD method outperforms the other methods (AFD, VCD, HOI, ESF, EDAEP, VDI, and MLRI) at 0.036, 0.026, 0.041, 0.005, 0.080, 0.064, and 0.046 on average, respectively. Table 4ZE results on the Kodak dataset for several demosaicking methods (the best performances are marked in bold).
3.2.Visual Performance AnalysisIn this section, we provide the subjective performance comparison between the BD method and the conventional methods. Figures 7Fig. 8–9 show the reconstructed images and their subjective performances in the visual comparison. Aliasing occurs near the image details because of signal information loss and poor restoration of high-frequency components. Therefore, the focus of the evaluation of the reconstruction of images should be the quality of the image edges. Figure 7 shows the zoomed demosaicking images on Kodak #8. From Fig. 7, we could see that our proposed algorithm has better visual performance at the white color reconstruction. One of evidences is that other conventional methods results show orange-blue color artifact. For example, the VDI method shows good visual quality seen in Fig. 7(g); however, compared with other subjective performances, VDI is much poorer than our algorithm, such as CPSNR, VDI is almost 4 dB less than our proposed algorithm. To evaluate subjective performance in high-frequency images, we adopted a color radial resolution chart.16 We tested eight methods on the TE216 and TE253 images, and the results are shown in Figs. 8 and 9. We noticed that the proposed method provides the best visual quality in high-frequency areas, whereas the conventional methods still showed orange-purple artifacts. We have requested 20 observers to assess the quality of the result images that are produced by proposed method and the conventional methods. All observers provide a score for each result image, from 0 to 10. If the result image is identical to the original image, score 10 is provided. If result image is poorer than the original image, lower score is provided. Table 5 shows average scores evaluated by the 20 observers. From Table 5, we could conclude that our proposed method has highest favor from the observers’ evaluation. Table 5Subjective performance assessment by observers’ evaluation.
3.3.Performance Comparison in the Noise ConditionNoise is introduced during image acquisition from the sensor. To evaluate the robustness of the proposed method, we conducted experiments on Kodak images contaminated by applying Gaussian and Poisson noises. Five selected conventional methods were used for comparison: VCD, HOI, ESF, EDAEP, and EDUSC. 3.3.1.Gaussian noise caseTable 6 shows the CPSNR results under the condition of Gaussian noise. Our approach showed the best CPSNR, providing a gain of 0.195 dB compared with the ESF. The proposed method obtained 5 out of 24 images in the second-order in the noise-free images and two images (#16 and #19) in the second-order in the Gaussian noisy images. Table 6CPSNR results on the Kodak dataset with Gaussian noise (σ=10, the best performances are marked in bold).
3.3.2.Poisson noise caseSome denoising approaches are based on the assumption that noise is Gaussian distributed. This assumption has been widely accepted, but recent research on denoising has considered noise as Poisson distributed given the improvement of the sensor with a reduced fixed pattern noise. The dominant noise type is the photon shot noise, which is strongly related to signal. Therefore, to verify the performance of the proposed algorithm in noisy images, we show the results for the Poisson noisy images of the Kodak dataset. Table 7 presents the CPSNR results on the demosaicking algorithms for Poisson noisy images. The proposed algorithm surpasses other conventional methods, and it improves images by 0.441 dB compared with the ESF. Similar to the Gaussian case, only 2 out of 24 images are ranked in the second-order for our proposed method, and the remaining ones are ranked in the first-order. Table 7CPSNR results on the Kodak dataset with Poisson noise (σ=10, the best performances are marked in bold).
3.4.Performance Test in a High-Frequency AreaTo compare visual performance in high-frequency images, we adopted two original high-frequency images: TE216 and TE253.16 We tested seven methods on both images shown in Figs. 6(b) and 6(c). Figures 8 and 9 show the reconstructed images of TE216 and TE253. The proposed method provides the best visual quality in the high-frequency area, followed by the MLRI method. It can be found from Fig. 8 that the VCD method shows the best visual quality in the high-frequency area, followed by the proposed method. For Fig. 9 image, the proposed method outperformed all the other benchmark methods. Although the proposed method provides the best objective performance, it is well known that the objective metrics (CPSNR, S-CIELAB, and FSIM) do not always rank quality of an image in the same way that observer does. There are many other factors considered by the human visual system and the brain, and therefore the perceived result could be different. 4.ConclusionIn this study, we proposed a demosaicking algorithm using an HBE in consideration of both pointwise and patchwise similarities. We introduced a Bayesian estimator to formulate the demosaicking model with MAP measurement. Our proposed method has superior performance in both noise-free and noisy images compared with conventional methods in both objective and subjective performances. AcknowledgmentsThis research was supported by Post-Doctor Research Program (2015) through the Incheon National University (INU), Incheon, South Korea. ReferencesB. Leung, G. Jeon and E. Dubois,
“Least-squares luma-chroma demultiplexing algorithm for Bayer demosaicking,”
IEEE Trans. Image Process., 20
(7), 1885
–1894
(2011). https://doi.org/10.1109/TIP.2011.2107524 IIPRE4 1057-7149 Google Scholar
G. Jeon and E. Dubois,
“Demosaicking of noisy Bayer-sampled color images with least-squares luma-chroma demultiplexing and noise level estimation,”
IEEE Trans. Image Process., 22
(1), 146
–156
(2013). https://doi.org/10.1109/TIP.2012.2214041 IIPRE4 1057-7149 Google Scholar
N. X. Lian et al.,
“Adaptive filtering for color filter array demosaicking,”
IEEE Trans. Image Process., 16
(10), 2515
–2525
(2007). https://doi.org/10.1109/TIP.2007.904459 IIPRE4 1057-7149 Google Scholar
K.-H. Chung and Y.-H. Chan,
“Color demosaicing using variance of color differences,”
IEEE Trans. Image Process., 15
(10), 2944
–2955
(2006). https://doi.org/10.1109/TIP.2006.877521 IIPRE4 1057-7149 Google Scholar
J. S. J. Li and S. Randhawa,
“Color filter array demosaicking using high-order interpolation techniques with a weighted median filter for sharp color edge preservation,”
IEEE Trans. Image Process., 18
(9), 1946
–1957
(2009). https://doi.org/10.1109/TIP.2009.2022291 IIPRE4 1057-7149 Google Scholar
I. Pekkucuksen and Y. Altunbasak,
“Edge strength filter based color filter array interpolation,”
IEEE Trans. Image Process., 21
(1), 393
–397
(2012). https://doi.org/10.1109/TIP.2011.2155073 IIPRE4 1057-7149 Google Scholar
L. Chang and Y.-P. Tam,
“Effective use of spatial and spectral correlations for color filter array demosaicing,”
IEEE Trans. Consum. Electron., 50
(1), 355
–365
(2004). https://doi.org/10.1109/TCE.2004.1277885 ITCEDA 0098-3063 Google Scholar
W.-J. Chen and P.-Y. Chang,
“Effective demosaicking algorithm based on edge property for color filter arrays,”
Digital Signal Process., 22
(1), 163
–169
(2012). https://doi.org/10.1016/j.dsp.2011.09.006 DSPREJ 1051-2004 Google Scholar
X. Chen, G. Jeon and J. Jeong,
“Voting-based directional interpolation method and its application to still color image demosaicking,”
IEEE Trans. Circuits Syst. Video Technol., 24
(2), 255
–262
(2014). https://doi.org/10.1109/TCSVT.2013.2255421 Google Scholar
C.-Y. Su and W.-C. Kao,
“Effective demosaicing using subband correlation,”
IEEE Trans. Consum. Electron., 55
(1), 199
–204
(2009). https://doi.org/10.1109/TCE.2009.4814435 ITCEDA 0098-3063 Google Scholar
D. Kiku et al.,
“Minimized-Laplacian residual interpolation for color image demosaicking,”
Proc. SPIE, 9023 90230L
(2014). https://doi.org/10.1117/12.2038425 PSISDG 0277-786X Google Scholar
F. Godtliebsen, E. Spjotvoll and J. S. Marron,
“A nonlinear Gaussian filter applied to images with discontinuities,”
J. Nonparametric Stat., 8 21
–43
(1997). https://doi.org/10.1080/10485259708832713 NOSTEK 1048-5252 Google Scholar
I. Pekkucuksen and Y. Altunbasak,
“Multiscale gradients-based color filter array interpolation,”
IEEE Trans. Image Process., 22
(1), 157
–165
(2013). https://doi.org/10.1109/TIP.2012.2210726 IIPRE4 1057-7149 Google Scholar
“Kodak lossless true color image suite,”
(2018) http://r0k.us/graphics/kodak/ April ). 2018). Google Scholar
X. Zhang and B. A. Wandell,
“A spatial extension of CIELAB for digital color image reproduction,”
J. Soc. Inf. Disp., 5
(1), 61
–63
(1997). https://doi.org/10.1889/1.1985127 JSIDE8 0734-1768 Google Scholar
L. Zhang et al.,
“FSIM: a feature similarity index for image quality assessment,”
IEEE Trans. Image Process., 20
(8), 2378
–2386
(2011). https://doi.org/10.1109/TIP.2011.2109730 IIPRE4 1057-7149 Google Scholar
BiographyJin Wang received her BS degree in mathematics and applied mathematics from Zhejiang University, Hangzhou, China, and her MS and PhD degrees in electronic communications engineering from Hanyang University, Seoul, Korea, in 2007 and 2013, respectively. Currently, she is an associate professor at Xidian University in Xi’an, Shaanxi, China. She has been a researcher in the Department of Electronics and Computer Engineering, Hanyang University and the Department of Embedded Systems Engineering, Incheon University. Jiaji Wu received his BS degree in electrical engineering from Xidian University, Xi’an, China, in 1996, his MS degree from National Time Service Center, the Chinese Academy of Sciences in 2002, and his PhD in electrical engineering from Xidian University in 2005. Currently, he is a professor at Xidian University, Xi’an China. His current research interests include image processing, still image coding, hyperspectral/multispectral image compression, communication, and high performance computing. Zhensen Wu received his BSc degree in applied physics from Xi’an Jiaotong University, Xi’an, China, in 1969, and his MSc degree in space physics from Wuhan University, Wuhan, China, in 1981. Currently, he is a professor at Xidian University, Xi’an, China. From 1995 to 2001, he was invited multiple times as a visiting professor to Rouen University, France, for implementing joint study of two projects supported by the Sino-France Program for Advanced Research. Marco Anisetti is an assistant professor at the Università degli Studi di Milano. His research interests are in the area of computational intelligence and its application to the design of complex systems and services. He investigates/has investigated the adoption of advanced techniques based on Kalman filters, rough sets, and fuzzy set theories for: (i) low-level functionalities like display management, advanced image processing and mobile geolocation, and (ii) high-level applications in the areas of humanized computing, human machine interaction and ambient intelligence. Gwanggil Jeon received his BS, MS, and PhD degrees from Hanyang University, in 2003, 2005, and 2008, respectively. From 2009 to 2011, he was a postdoctoral fellow at University of Ottawa, and from 2011 to 2012, he was an assistant professor at Niigata University. Currently, he is a professor at Incheon National University. His research interests fall under the umbrella of image processing, particularly image compression, motion estimation, demosaicking, and image enhancement as well as computational intelligence such as fuzzy and rough sets theories. |