Open Access
9 May 2018 Bayesian method application for color demosaicking
Jin Wang, Jiaji Wu, Zhensen Wu, Marco Anisetti, Gwanggil Jeon
Author Affiliations +
Abstract
This study presents a Bayesian approach based on a color image demosaicking algorithm. The proposed method is composed of pointwise and patchwise measurements. The estimation of the missing pixel is formulated as a maximum a posteriori and a minimum energy function. By utilizing Bayesian theory and some prior knowledge, the missing color information is estimated with a statistics-based approach. Under the maximum a posteriori and Bayesian framework, the desired target image corresponds to the optimal reconstruction given the mosaicked image. Compared with existing demosaicking methods, the proposed algorithm improves the CPSNR, S-CIELAB, FSIM, and zipper effect measurements while maintaining high efficiency. Moreover, it handles Gaussian and Poisson noisy images better than other conventional images.

1.

Introduction

To reduce cost, when a digital camera captures a color image, only one color channel out of three color channels is captured. The single charge-coupled device sensor estimates the amount of red, green, and blue in the pixel. This sensor is covered with a color filter array (CFA), which causes each pixel in the camera to capture only one color channel. As pixels only have single color information, the other two color components are reconstructed, and this process is called demosaicking. One of the most famous CFA patterns called Bayer CFA is shown in Fig. 1. In this pattern, each group has nine pixels, in which five are green pixels located in a quincunx manner, and the other four pixels are red and green pixels.

Fig. 1

7×7 CFA Bayer pattern.

OE_57_5_053102_f001.png

Based on the Bayer pattern, various demosaicking methods were proposed in the past few decades.1,2 These methods were developed to acquire a full-color image by utilizing the color differences among three color planes in the spatial or frequency domains. Especially, the adaptive color plane interpolation is an early technique that uses an edge sensing interpolation.3 This method uses three predictors and then chooses one of them via an edge classifier using the Laplacian second-order derivatives for the color components and the gradient term for the green component. Based on this approach, several methods were proposed. The adaptive filtering for CFA demosaicking (AFD) uses frequency analysis of Bayer CFA samples and tries to extract high-frequency information by a properly designed filter.4 Based on this analysis, AFD applies a low-pass filter for the luminance of green positions to better reconstruct high-frequency information. The high-frequency information at the red/blue positions is then reconstructed by adaptive filtering of color difference components. Color interpolation using the variance of color differences (VCD) is a heuristic approach used to determine a distinct edge;5 in this approach, the missing green values are grouped into plane and edge regions. This method uses the heuristic demosaicking approach. The high-order interpolation (HOI) method is based on a Taylor series that uses a weighted median filter to select a predictor.6 This method consists of two steps: (1) interpolation to determine four estimates of a missing color and (2) an edge orientation map-based classifier. In Ref. 7, the authors proposed a 3×3 edge strength filter that provides local, orientation-free luminance transition information. The edge strength filter is useful in recognizing the areas where a constant color difference assumption may fail, thus improving the demosaicking performance. By adding fixed weights to Ref. 8, an effective demosaicking method based on the edge property (EDAEP) was proposed.9 The authors used an accurate edge detecting method to general different weights to minimize false colors in the edge areas. In Ref. 10, authors proposed to combine voting-based directional interpolation (VDI) with weighting-based interpolation for a more effective interpolation without estimating the interpolation directions erroneously in a complex region. The VDI is based on a gradient-inverse weighted interpolation along the interpolation direction as determined by a voting approach. The effective demosaicking based on subband correlation (EDUSC) was proposed using a discrete wavelet transform to discern edge pixels.11 The EDUSC method uses a subimage subband correlation to enable a good initial interpolation and accurate edge detection. The EDUSC method has three parts: (1) initial interpolation in each color channel, (2) missing pixel calculation from two other color channels, and (3) iterative subband synthesis for artifact reduction. In minimized-Laplacian residual interpolation (MLRI),12 the authors estimated the tentative pixel values by minimizing the Laplacian energies of the residuals. Then, the authors incorporated the MLRI into the gradient-based threshold-free algorithm to restore the demosaicking results more precisely.

In this study, we propose a Bayesian approach-based demosaicking algorithm (BD) with a hybrid method in consideration of pointwise and pathwise similarity. The demosaicking algorithm recovers the underlying original true color image from a mosaicked image. We formulate this method as a maximum a posteriori (MAP) problem. Some statistical justification should be investigated to solve the MAP problem. Therefore, we present a statistical interpretation and a Bayesian motivation to clarify the MAP issue. The MAP problem focuses on the distribution of the residual value between the missing color pixel and its neighboring color pixels. We examine the pointwise and patchwise models: the pointwise model preserves a large-scale structure, and the patchwise model provides a geometric outline and a fine-scale structure in a large search region. In the computational complexity concept, the pointwise model is simpler and more efficient than the patchwise model. To utilize the different advantages of point and patch structures, we design an adaptive hybrid model as a linear combination of the two models, and the parameter is adaptive to a standard deviation representing the structural feature. After the interpolation of the missing green channel, we estimate the red/blue channel by utilizing the uniform color difference in a local region. Afterward, we apply demosaicking algorithms to the Gaussian and Poisson noisy images to prove the robustness of the proposed algorithm.

The rest of the paper is organized as follows: in Sec. 2, we describe the proposed algorithm. In Sec. 3, some experimental results and the corresponding discussion are provided. Finally, the conclusion is drawn in Sec. 4.

2.

Proposed Algorithm

2.1.

Bayesian Approach-Based Analysis

Let a mosaicked image be x(k)=[xR(k)xG(k)xB(k)], where k=(i,j)R2 is the image coordinate index. We assume that the true original color image y=[yR(k)yG(k)yB(k)] is highly related to x, as shown in Fig. 2. An estimator y^ is considered a Bayes estimator if it minimizes E[L(y,y^)|x] among all estimators. It is equivalent to an MAP solution, in which L(y,y^) is a loss function used as some function of the difference between estimated and true values. The quadratic loss function L(y,y^)=y^y2 is commonly used in signal processing, and λ(y) is a constant. The optimal Bayesian estimator (OBE) y^OBE is determined by minimizing E[L(y,y^)|x] as

Eq. (1)

y^OBE=argminy^E[y^y2|x].

Fig. 2

Observation of the low resolution for the Bayer CFA.

OE_57_5_053102_f002.png

Based on the assumption of the quadratic loss function,13 the OBE is

Eq. (2)

y^OBE=E[y|x]=yΛyp(y|x)=yΛyp(xy)p(x),
where Λ is the large space of all configurations of y (|Λ|=256n if y(x){0,,255}). Here parameter n is the size of the data. By using the Bayes rules and the law of total probability, we have
p(x,y)=p(x|y)p(y),p(x)=yΛp(x|y)(y).

Then, we can modify Eq. (2) into

Eq. (3)

y^OBE=yΛyp(x|y)(y)yΛp(x|y)(y).

As shown in Eq. (3), to obtain the optimal estimation y^OBE, we determine the distribution of x|y and the distribution of y known as a prior distribution. To derive the distributions, we need to utilize the observation in a local neighbor region Sy centered at the to-be-interpolated pixel instead of the pixels at the same position. In the local region Sy, we assume that the prior distribution of y is a uniform distribution, that is, p(y)=1/|Sy|. In this local neighbor region, we assume that the observation has similar structures. To achieve the optimal estimation, we use the following approximation:13

Eq. (4)

1|Sy|mSyy(k+m)p[x(k)|y(k+m)]yΛyp(x|y)p(y),1|Sy|mSyp[x(k)|y(k+m)]yΛp(x|y)p(y).

Then, Eq. (3) is changed into

Eq. (5)

y^OBE1|Sy|mSyy(k+m)p[x(k)|y(k+m)]1|Sy|mSyp[x(k)|y(k+m)],
where k+m can be any neighbor pixels in a square region (2ws+1)×(2ws+1). In this paper, we choose ws=1 for high efficiency. The target is changed to find the distribution of x(k)|y(k+m). Figures 3(b)3(d) show that it does not have the obvious statistical property for all the three color channels. Based on the histogram of residual images, we derive the distribution of x(k)y(k+m)|y(k+m) instead of x(k)|y(k+m). Figures 3(e)3(g) indicate that x(k)y(k+m)|y(k+m) has a clear statistical feature. Two well-known distributions, namely, Gaussian and Laplacian, have this kind of form. We compare these two kinds of distributions with different parameters in Fig. 4. The Laplacian distribution fits the statistical property better than the Gaussian distribution. The distribution of x(k)|y(k+m) follows the Laplacian distribution.

Fig. 3

Illustration of the image statistic model. (a) Original image, (b) p[x(k)|y(k+m)] in the R color channel, (c) p[x(k)|y(k+m)] in the G color channel, (d) p[x(k)|y(k+m)] in the B color channel, (e) p[x(k)y(k+m)|y(k+m)]; in the R color channel, (f) p[x(k)y(k+m)|y(k+m)] in the G color channel, and (g) p[x(k)y(k+m)|y(k+m)] in the B color channel.

OE_57_5_053102_f003.png

Fig. 4

Matching of the statistical model and empirical histogram with different parameters of the Gaussian and Laplacian distributions. (a) f(x)=3.8×105exp(0.026x2), (b) f(x)=4.6×105exp(0.16|x|), (c) f(x)=2.8×105exp(0.0081x2), (d) f(x)=4.2×105exp(0.137|x|), (e) f(x)=1.7×105exp(0.0028x2), and (f) f(x)=3×105exp(0.096|x|).

OE_57_5_053102_f004.png

Equation (5) is modified into

Eq. (6)

y^G(k)=mSyy˜G(k+m)C(y)mSyC(y),
where C(y)=exp[vG(k)vG(k+m)1/α].

The weight parameter α plays an important role in the statistic model. For a different weight parameter α, the matching has varying features, as shown on the left in Fig. 4. In the figure, we depict the empirical histogram of residual images and compare the empirical histogram with the prior statistical model of varying parameters. From the top-left to the bottom-left, the weight parameter α is 0.35, 0.36, and 0.4. The top one represents the parameters with the smallest α, which achieves good matching at zero but fails at tails. The bottom one represents the parameters with the largest α, which achieves good matching at tails but fails at zero. In Fig. 4, the width plays a more important role than height because when the residual is zero, this pixel is in the smooth region, which is easy to be reconstructed with conventional methods. In terms of image restoration, the focus is on the complex region such as edges or textures, that is, the tails of the distribution. The tails are controlled by the width. To achieve the best matching with the empirical histogram, we should assign adaptive weight parameters to different pixel values. That is, large residual values should have large weight parameters, and small residual values should have small ones. Therefore, we design an adaptive weight parameter that is proportional to the variance of the to-be-interpolated pixel in a local region

Eq. (7)

α=λσm=mSyλx(k+m)x¯|Sy|,
where x is the expected value in the local region Sy with size (2ws+1)×(2ws+1). Through Eq. (6), we can determine the original image y(k+m) to obtain the optimal estimation of the color image. The original image is partially unknown, but it is also partially given in the mosaicked image. Thus, in Eq. (6), we substitute the unknown y(k+m) into the estimated direction value of the given mosaicked color pixel values.

2.2.

Green Channel Reconstruction Using a Bayesian Framework

According to the discussion in Sec. 2.1, we first restore the green channel and then reconstruct the red/blue channel based on the entire green channel and the given mosaicking color pixels. Figure 1 shows that a missing green color pixel at a given red/blue color pixel is surrounded by the four nearest given green color pixels in the north, south, west, and east directions. We still have the red/blue color pixel in the missing green color pixel. Therefore, we should consider this feature to reconstruct the missing green color pixel. The mosaicked image has a special feature that has pixel values in all locations. Based on this specific property of mosaicking image, instead of directly using the green color pixels in the four directions in Eq. (6), we adopt the Taylor approximation14 to provide the approximated four-direction color values by the given color pixel values

Eq. (8)

yG(k+m)=xG(k+m)+xR(k)xR(k+2m)2+xR(k+3m)+xR(km)2xR(k+m)8.

In Eq. (6), we have two candidates for distance x(k)y(k+m), namely, pointwise and patchwise. The patchwise distance aims to preserve the structure of the images in a large local region. In a small search region, especially in terms of image restoration, pointwise similarity plays an important role in addressing a large-scale structure. Both pointwise and patchwise distances should be considered to estimate the missing color pixels. Therefore, we propose the following hybrid Bayesian estimator (HBE)-BD algorithm:

Eq. (9)

C(y)=exp[vG(k)vG(k+m)12α]+μexp[xR(k)xR(k+m)1α],
where vG(k) and vG(k+m) are the patch vectors containing the given center red pixel and four neighboring green pixels. The neighboring region Sy is shown in Fig. 5. In this algorithm, μ=0.1.

Fig. 5

Structure of the neighboring region. (a) SyH is the region containing the given color pixels in the horizontal direction and with size 1×(2ws+1), (b) SyV is the region containing the given color pixels in the vertical direction and with size (2ws+1)×1, and (c) SyRB contains the four given red pixels with size (2ws+1)×(2ws+1).

OE_57_5_053102_f005.png

For images with complex features, such as various directions or textures, Eq. (6) can obtain good performance. For the two special cases of horizontal and vertical edges, we should apply the given pixels in the horizontal or vertical direction to reconstruct the missing color pixel. Horizontal and vertical directions exist in most images and occupy a large proportion in one image. Therefore, we estimate the two directions by modifying Eq. (6) and adopting the horizontal and vertical components as follows:

Eq. (10)

y^GH(k)=mSyHy˜G(k+m)C(y)mSyHC(y),y^GV(k)=mSyVy˜G(k+m)C(y)mSyVC(y),
where SyH and SyV are the regions containing the given color pixels in the horizontal and vertical directions and with size 1×(2ws+1) and (2ws+1)×1 (Fig. 5), respectively. Here, C(y) is adopted in Eq. (9). Then, a criterion is designed to distinguish the direction as follows:

Eq. (11)

y¯¯G(k)={y^GH(k),if  δ(k)>T1y^GV(k),if  δ(k)<T1y^G(k),otherwise,δ(k)={1,if  min(DH(k),DV(k),T2)=DH(k)1,if  min(DH(k),DV(k),T2)=DV(k)0,otherwise,DH=x*f+(xy^GH)*f,DV=x*f+(xy^GV)*f,
where T is the threshold, and DH and DV are normalized. The interpolated full green channel is then obtained.

2.3.

Red/Blue Plane Reconstruction

As we fully obtained a green plane in Sec. 2.2, we now populate red/blue pixels in blue/red or green locations. In the reconstruction of the red/blue color channel, we use the restored green color channel because of the strong correlation among the color channels. To reconstruct the missing red pixels in the given blue positions, four given red pixels are located around the given blue pixel (Fig. 1). We use the color difference based on the reconstructed green pixels as follows:

Eq. (12)

y¯R(k)=y¯G(k)+mSyRBΔ(k+m)C(y)mSyRBC(y),Δ(k+m)=xR(k+m)y¯G(k+m),C(y)=exp[vR(k)vR(k+m)1β],+μexp[y¯G(k)y¯R(k+m)1β],β=α2,
where vR(k) and vR(k+m) are the patch vectors including the four given red pixels located in the northwest, northeast, southwest, and southeast, and SyRB contains the same four given red pixels with size (2ws+1)×(2ws+1) as shown in Fig. 5(c).

For the missing red/blue pixel in the given G pixel, Eq. (13) is applied as shown in Fig. 1.

Eq. (13)

y¯R(k)=y¯G(k)+mSyRGΔ(k+m)C(y)mSyRGC(y),Δ(k+m)=xR(k+m)y¯G(k+m),C(y)=exp[y¯G(k)y¯R(k+m)1β],
where SyRG contains the two given red pixels above and below the given green pixel.

2.4.

Refinement of Each Color Channel

Once the green channel is populated using Eq. (11), for the given red pixels, the green channel can be further refined as

Eq. (14)

y¯R(k)=xR(k)+κ[y^G(k)xR(k)]+(1κ)[mS(x)]Δ(k+m)x^G(k+m)[mS(x)]Δ(k+m),x^G(k+m)=xG(k+m)x¯R(k+m),Δ(k+m)={|y^G(k)xG(k+m)|+a{0,2}|xR(k+am)xR[k+(a+2)m]|}1,
where κ is the weight parameter assumed to be 0.1. Then, we refine the red/blue channel by the modified green channel as in Sec. 2.2. We iteratively refine the green channel and red/blue channel by using the above procedure until we obtain the optimal results or reach the stop condition.

3.

Experimental Results

This section presents the detailed performance comparison. Simulations were conducted on 512×768 size Kodak dataset15 and two high-frequency images16 to assess the superiority of the proposed method compared with the conventional methods. The test images are shown in Fig. 6. Eight methods were selected as benchmarks: AFD,4 VCD,5 HOI,6 ESF,7 EDAEP,8 VDI,10 EDUSC,11 and MLRI.12 We evaluated the objective performance in terms of CPSNR, S-CIELAB ΔE*, FSIM, and the zipper effect (ZE) aside from the visual performance. Afterward, we apply the Gaussian and Poisson noise to the Kodak dataset and provide the CPSNR results through a comparison with the existing demosaicking methods. Our simulation was conducted on an Intel Core 2 Duo CPU E8500 @ 3.16 GHz. The window size ws was set to 2.

Fig. 6

(a) 24 images of the Kodak dataset, (b) high-frequency image TE216, and (c) high-frequency image TE253.

OE_57_5_053102_f006.png

3.1.

Objective Performance Analysis

We first compare the objective performances measured in terms of CPSNR. Table 1 displays the comparison of the CPSNR results of the presented method with those of other benchmark methods. Seven methods were selected as benchmarks: AFD, VCD, HOI, ESF, EDAEP, VDI, and MLRI. As shown in Table 1, the proposed method has the best average CPSNR, and the gain are 1.229 (AFD), 0.916 (VCD), 1.473 (HOI), 0.519 (ESF), 2.786 (EDAEP), 2.699 (VDI), and 1.233 (MLRI). Although images 4, 15, 16, 23, and 24 did not show the best CPSNR using the proposed method, our method still ranks second. “Diff.” shows the CPSNR difference between BD and the corresponding method. “Rank” indicates the rank of the BD method compared with those of the benchmark methods. The best performances are marked in bold.

Table 1

CPSNR performance comparison on the Kodak images (the best performances are marked in bold).

ImageAFDVCDHOIESFEDAEPVDIMLRIBDRank
137.55838.41936.56439.90934.91935.34336.80540.6761
240.740.43340.82740.82139.73639.60740.79141.5651
342.68642.53742.7142.55441.74941.31842.98743.5171
441.05540.76940.71640.44740.03439.5641.4941.3632
538.03937.78337.63437.5535.78236.54237.65738.681
638.03639.99938.80741.21736.84536.9839.13141.3861
742.90742.13342.72242.15441.28141.82842.84143.2231
835.2336.28934.85537.18932.83233.99534.89437.6311
942.5742.8542.39142.95940.90241.68742.37743.4391
1042.69942.46542.26842.60240.97841.16242.41943.1541
1139.33839.88939.12840.69337.47637.68539.30540.9261
1242.77443.4343.02743.80141.66941.70543.26844.2771
1333.76934.79832.76736.10631.48431.23333.14136.3261
1437.16637.01237.17336.11336.26836.33637.61537.8981
1539.83939.69539.05439.27438.80538.07139.33939.8082
1641.14843.71342.46744.77240.44140.6542.87944.2892
1741.39241.51940.85241.95339.63139.63940.98542.1071
1837.11837.11636.44937.7235.37934.96736.58537.861
194040.50639.8241.49437.75238.86939.95441.6841
2041.11741.12840.62941.52339.2639.74440.68141.7521
2138.67639.1637.91540.29736.45236.97138.18140.3751
2238.51238.07238.03138.40937.17337.12338.61438.9651
2343.14942.99543.48342.44942.35641.30943.8143.7262
2434.83235.10334.16235.34233.72832.70834.70435.1772
Avg.39.59639.90939.35240.30638.03938.12639.60240.8251
Diff.1.2290.9161.4730.5192.7862.6991.223

To numerically evaluate the results of visual performance, we used the S-CIELAB ΔE* metric.17 S-CIELAB ΔE* metric is a spatial form of the CIELAB ΔE* developed for determining the distance between the S-CIELAB ΔE* representation of an original and that of the reconstructed image. In terms of the S-CIELAB ΔE* measure in the Kodak dataset, the BD method provides higher performance on average than all the other methods (AFD, VCD, HOI, ESF, EDAEP, VDI, and MLRI) with factors of 0.0522, 0.0450, 0.0703, 0.0337, 0.1789, 0.1319, and 0.0612, respectively, as shown in Table 2.

Table 2

S-CIELAB ΔE* results on Kodak dataset for several demosaicking methods (the best performances are marked in bold).

ImageAFDVCDHOIESFEDAEPVDIMLRIBDRank
11.19891.12461.25050.99541.47361.3681.20320.96481
20.64270.68180.64420.66340.70290.72830.66150.60611
30.48890.5010.49160.50950.52430.54130.47060.46321
40.67210.69070.70030.72660.72850.79530.64060.65952
51.01361.05611.06411.24191.27061.10741.11071.04062
60.95710.81250.85150.74130.98890.94980.83470.7552
70.52290.59420.52770.58380.66370.56410.53640.50811
81.41271.28631.41581.19041.82451.47731.45971.17271
90.5820.57470.60070.58520.65290.62380.57720.54851
100.53670.55630.57150.5690.61020.6170.55540.53271
110.80320.7630.78030.72230.92260.85870.76950.70971
120.50620.47790.48370.46550.55010.5290.47940.44051
131.69591.64111.8581.47712.01242.04171.82161.54822
141.01790.9950.98731.04331.13761.09140.93640.88941
150.64230.67120.70070.72150.69070.72570.68290.66732
160.69830.55560.59790.5080.70420.68920.56280.58284
170.54190.55290.5640.54260.64090.61750.57130.53651
181.00771.05751.06941.09191.161.14661.05181.05633
190.78210.77460.80360.73570.95650.84660.78530.73431
200.54590.56210.57640.55530.64740.57850.6750.54581
210.92910.91720.98860.86061.09051.01890.9620.89172
220.94211.03440.98051.02291.05630.99520.9140.92432
230.50090.5280.48770.55960.53590.54390.47090.48722
241.01811.07751.09921.10171.15611.11831.1431.14166
Avg.0.81910.81190.83730.80060.94590.89890.82820.7671
Diff.0.05220.0450.07030.03370.17890.13190.0612

To further evaluate the performance of the BD method, we adopted a third metric called the FSIM index.18 The best FSIM result is 1.0, whereas a smaller FSIM indicates poor visual quality. Table 3 presents the numerical FSIM results on the Kodak dataset for various benchmarks. The BD method outperforms the other methods (AFD, VCD, HOI, ESF, EDAEP, VDI, and MLRI) at 0.0003, 0.0002, 0.0006, 0.0001, 0.0009, 0.0010, and 0.0003 on average, respectively.

Table 3

FSIM results on the Kodak dataset for several demosaicking methods (the best performances are marked in bold).

ImageAFDVCDHOIESFEDAEPVDIMLRIBDRank
10.99860.99870.9980.9990.99740.99750.99830.99921
20.99850.99840.99820.99840.99810.99760.99830.99861
30.99930.99920.99920.99920.99910.99860.99930.99941
40.99920.99920.99890.99920.99890.99860.99920.99921
50.99890.99880.99880.99880.99820.99850.99880.99911
60.99870.9990.99870.99930.99820.99820.99890.99931
70.99950.99950.99950.99950.99930.99930.99950.99961
80.99780.99820.99750.99850.99630.99720.99770.99871
90.99920.99920.9990.99930.99880.99880.99910.99931
100.99930.99920.99910.99930.99890.99880.99920.99931
110.99880.99890.99860.99910.99810.99810.99880.99911
120.99910.99910.9990.99930.99870.99870.99910.99931
130.99740.99780.99620.99850.99540.99530.99690.99871
140.99870.99870.99850.99850.99830.99790.99870.99891
150.99860.99870.99780.99850.9980.99720.99830.99844
160.9990.99940.99910.99950.99880.99880.99930.99942
170.99930.99930.99910.99940.99890.9990.99930.99941
180.99830.99820.99770.99860.99730.9970.99780.99861
190.9990.99910.99870.99920.99840.99860.9990.99931
200.99910.9990.99890.99910.99870.99870.99910.99921
210.99850.99860.99810.99890.99760.99780.99840.99891
220.99840.99820.99810.99830.99780.99780.99850.99851
230.99930.99930.99930.99920.99920.9990.99940.99941
240.99750.99750.99680.99770.99660.99590.99730.99762
Avg.0.99880.99880.99850.99890.99810.9980.99870.9991
Diff.0.00030.00020.00060.00010.00090.0010.0003

The ZE is one of the annoying artifacts in demosaicking. Lu and Tan introduced a measurement to evaluate the existence of the ZE. The result of the ZE represents the percentage of the pixel containing the ZE. Thus, the smaller the number is, the fewer the images containing the ZE. Table 4 shows that the BD method generated the smallest ZE. The BD method outperforms the other methods (AFD, VCD, HOI, ESF, EDAEP, VDI, and MLRI) at 0.036, 0.026, 0.041, 0.005, 0.080, 0.064, and 0.046 on average, respectively.

Table 4

ZE results on the Kodak dataset for several demosaicking methods (the best performances are marked in bold).

ImageAFDVCDHOIESFEDAEPVDIMLRIBDRank
10.3840.3480.3960.2990.4610.4420.4160.2761
20.2410.2410.2390.2320.2680.2570.240.2141
30.1280.1220.1270.1110.1460.1420.1280.1061
40.2230.2290.2330.2180.260.2590.2260.2091
50.3080.3140.3070.2940.3830.3380.3240.2781
60.2970.2470.2660.2060.3220.3120.2680.2072
70.1450.1580.1470.1490.1810.1580.1530.1341
80.4030.3730.4010.3380.4760.4220.4250.3261
90.1810.1720.20.160.2290.2210.2180.1642
100.1710.1720.1910.1610.2180.210.20.161
110.250.2310.2380.1940.2820.2730.240.1931
120.1510.1340.150.1140.1760.1690.1510.1172
130.4750.450.4950.4080.5380.5370.5040.3971
140.2820.2610.2770.2340.3370.3160.2890.2341
150.2120.2130.2260.2040.2450.2380.230.2031
160.2090.1570.1810.1280.2230.2130.1780.1512
170.1740.1710.1890.1560.2270.2150.1930.1531
180.3460.3520.3670.3410.40.3920.3710.3341
190.2730.2540.2850.2330.3350.3040.290.2311
200.2430.2460.2530.2320.2740.2590.2480.2311
210.2850.2730.2990.2440.3350.3180.3080.2462
220.2980.3160.3150.2990.3390.320.3070.2861
230.1540.1660.1610.160.1730.1650.1610.1511
240.2590.2610.270.2410.3190.2840.2810.2371
Avg.0.2540.2440.2590.2230.2980.2820.2650.2181
Diff.0.0360.0260.0410.0050.080.0640.046

3.2.

Visual Performance Analysis

In this section, we provide the subjective performance comparison between the BD method and the conventional methods. Figures 7Fig. 89 show the reconstructed images and their subjective performances in the visual comparison. Aliasing occurs near the image details because of signal information loss and poor restoration of high-frequency components. Therefore, the focus of the evaluation of the reconstruction of images should be the quality of the image edges. Figure 7 shows the zoomed demosaicking images on Kodak #8. From Fig. 7, we could see that our proposed algorithm has better visual performance at the white color reconstruction. One of evidences is that other conventional methods results show orange-blue color artifact. For example, the VDI method shows good visual quality seen in Fig. 7(g); however, compared with other subjective performances, VDI is much poorer than our algorithm, such as CPSNR, VDI is almost 4 dB less than our proposed algorithm.

Fig. 7

(a) Part of the original Kodak #8 image. Perceived image quality comparison using various deinterlacing methods: (b) AFD, (c) VCD, (d) HOI, (e) ESF, (f) EDAEP, (g) VDI, (h) MLRI, and (i) the proposed method.

OE_57_5_053102_f007.png

Fig. 8

(a) Original TE216 image. Perceived image quality comparison using various deinterlacing methods: (b) AFD, (c) VCD, (d) HOI, (e) ESF, (f) EDAEP, (g) VDI, (h) MLRI, and (i) the proposed method.

OE_57_5_053102_f008.png

Fig. 9

(a) Original TE253 image. Perceived image quality comparison using various deinterlacing methods: (b) AFD, (c) VCD, (d) HOI, (e) ESF, (f) EDAEP, (g) VDI, (h) MLRI, and (i) the proposed method.

OE_57_5_053102_f009.png

To evaluate subjective performance in high-frequency images, we adopted a color radial resolution chart.16 We tested eight methods on the TE216 and TE253 images, and the results are shown in Figs. 8 and 9. We noticed that the proposed method provides the best visual quality in high-frequency areas, whereas the conventional methods still showed orange-purple artifacts.

We have requested 20 observers to assess the quality of the result images that are produced by proposed method and the conventional methods. All observers provide a score for each result image, from 0 to 10. If the result image is identical to the original image, score 10 is provided. If result image is poorer than the original image, lower score is provided. Table 5 shows average scores evaluated by the 20 observers. From Table 5, we could conclude that our proposed method has highest favor from the observers’ evaluation.

Table 5

Subjective performance assessment by observers’ evaluation.

MethodVCDHOIESFEDAEPEDUSCBDRank
Score9.20989.24049.1189.30169.34759.45461

3.3.

Performance Comparison in the Noise Condition

Noise is introduced during image acquisition from the sensor. To evaluate the robustness of the proposed method, we conducted experiments on Kodak images contaminated by applying Gaussian and Poisson noises. Five selected conventional methods were used for comparison: VCD, HOI, ESF, EDAEP, and EDUSC.

3.3.1.

Gaussian noise case

Table 6 shows the CPSNR results under the condition of Gaussian noise. Our approach showed the best CPSNR, providing a gain of 0.195 dB compared with the ESF. The proposed method obtained 5 out of 24 images in the second-order in the noise-free images and two images (#16 and #19) in the second-order in the Gaussian noisy images.

Table 6

CPSNR results on the Kodak dataset with Gaussian noise (σ=10, the best performances are marked in bold).

VCDHOIESFEDAEPEDUSCBDRank
132.86132.37933.24731.69432.70933.3281
233.37933.62933.45733.44333.48533.7551
333.79233.98633.79433.89733.89834.0711
433.46533.65433.40933.56233.56433.6581
532.69232.81732.63632.16432.84833.1161
633.29733.18633.63432.57232.91333.6661
733.67233.98133.69933.79333.85933.9811
832.1731.63932.58330.58431.65432.7311
933.80133.92733.83633.73733.84234.0451
1033.75433.91233.79133.74433.84734.0361
1133.23733.23133.44432.75433.22533.5871
1233.90634.0533.97633.87533.89734.1981
1331.55630.56632.14229.76731.29132.1511
1432.43532.66332.11432.3432.63532.8761
1533.47133.47733.37533.46833.50933.6581
1633.82333.86333.99233.54233.59233.7054
1733.62633.67433.70133.46933.71133.9051
1832.51732.43432.7431.99332.68232.9071
1933.42633.38133.55532.86333.31933.5422
2034.37134.43134.45934.10534.48434.6441
2133.12632.92733.40832.44333.11633.4211
2232.82832.99232.92732.73332.98533.2611
2333.85634.11833.78734.03733.97334.1351
2431.70731.39131.83931.16931.63931.851
Avg.33.19933.1833.31432.82333.19533.5091
Diff.0.3110.330.1950.6870.315

3.3.2.

Poisson noise case

Some denoising approaches are based on the assumption that noise is Gaussian distributed. This assumption has been widely accepted, but recent research on denoising has considered noise as Poisson distributed given the improvement of the sensor with a reduced fixed pattern noise. The dominant noise type is the photon shot noise, which is strongly related to signal. Therefore, to verify the performance of the proposed algorithm in noisy images, we show the results for the Poisson noisy images of the Kodak dataset. Table 7 presents the CPSNR results on the demosaicking algorithms for Poisson noisy images. The proposed algorithm surpasses other conventional methods, and it improves images by 0.441 dB compared with the ESF. Similar to the Gaussian case, only 2 out of 24 images are ranked in the second-order for our proposed method, and the remaining ones are ranked in the first-order.

Table 7

CPSNR results on the Kodak dataset with Poisson noise (σ=10, the best performances are marked in bold).

VCDHOIESFEDAEPEDUSCBDRank
137.94636.25539.25734.70337.3739.8891
239.90540.28240.24939.30440.0440.9211
341.44741.6441.48140.87741.63742.2621
440.04440.04739.78939.45940.09840.5021
537.44137.33337.23835.58237.67438.3061
639.20238.22340.22636.45337.50240.3461
741.04141.56941.07640.43841.6141.9331
835.9434.59136.77532.67934.78437.181
941.38841.10441.47139.98841.22141.8491
1041.18441.10441.30340.10241.28241.7471
1139.29738.64140.00637.15438.92140.2091
1241.51241.3541.80340.35941.05542.1531
1334.56932.63135.81431.38833.96136.0141
1436.6836.84535.85736.01937.05237.5171
1539.13438.58738.75138.35638.89939.2481
1642.26541.38443.01139.74240.30242.512
1740.80740.25941.1739.19440.92241.3231
1836.85336.25537.4535.21937.13337.5751
1939.66239.14540.47837.32639.33540.6431
2040.15939.78540.45538.63740.33840.691
2138.54137.4739.52136.12738.31739.5891
2237.59337.58237.90536.81237.88538.4261
2341.73742.17441.33241.33541.86642.2991
2434.86133.97335.07133.55934.54634.932
Avg.39.13438.67639.47937.53438.90639.9191
Diff.0.7861.2430.4412.3851.013

3.4.

Performance Test in a High-Frequency Area

To compare visual performance in high-frequency images, we adopted two original high-frequency images: TE216 and TE253.16 We tested seven methods on both images shown in Figs. 6(b) and 6(c). Figures 8 and 9 show the reconstructed images of TE216 and TE253. The proposed method provides the best visual quality in the high-frequency area, followed by the MLRI method. It can be found from Fig. 8 that the VCD method shows the best visual quality in the high-frequency area, followed by the proposed method. For Fig. 9 image, the proposed method outperformed all the other benchmark methods. Although the proposed method provides the best objective performance, it is well known that the objective metrics (CPSNR, S-CIELAB, and FSIM) do not always rank quality of an image in the same way that observer does. There are many other factors considered by the human visual system and the brain, and therefore the perceived result could be different.

4.

Conclusion

In this study, we proposed a demosaicking algorithm using an HBE in consideration of both pointwise and patchwise similarities. We introduced a Bayesian estimator to formulate the demosaicking model with MAP measurement. Our proposed method has superior performance in both noise-free and noisy images compared with conventional methods in both objective and subjective performances.

Acknowledgments

This research was supported by Post-Doctor Research Program (2015) through the Incheon National University (INU), Incheon, South Korea.

References

1. 

B. Leung, G. Jeon and E. Dubois, “Least-squares luma-chroma demultiplexing algorithm for Bayer demosaicking,” IEEE Trans. Image Process., 20 (7), 1885 –1894 (2011). https://doi.org/10.1109/TIP.2011.2107524 IIPRE4 1057-7149 Google Scholar

2. 

G. Jeon and E. Dubois, “Demosaicking of noisy Bayer-sampled color images with least-squares luma-chroma demultiplexing and noise level estimation,” IEEE Trans. Image Process., 22 (1), 146 –156 (2013). https://doi.org/10.1109/TIP.2012.2214041 IIPRE4 1057-7149 Google Scholar

3. 

J. F. Hamilton and J. E. Adams, “Adaptive color plane interpolation in single sensor color electronic camera,” U.S. Patent 5,629,734 (1997).

4. 

N. X. Lian et al., “Adaptive filtering for color filter array demosaicking,” IEEE Trans. Image Process., 16 (10), 2515 –2525 (2007). https://doi.org/10.1109/TIP.2007.904459 IIPRE4 1057-7149 Google Scholar

5. 

K.-H. Chung and Y.-H. Chan, “Color demosaicing using variance of color differences,” IEEE Trans. Image Process., 15 (10), 2944 –2955 (2006). https://doi.org/10.1109/TIP.2006.877521 IIPRE4 1057-7149 Google Scholar

6. 

J. S. J. Li and S. Randhawa, “Color filter array demosaicking using high-order interpolation techniques with a weighted median filter for sharp color edge preservation,” IEEE Trans. Image Process., 18 (9), 1946 –1957 (2009). https://doi.org/10.1109/TIP.2009.2022291 IIPRE4 1057-7149 Google Scholar

7. 

I. Pekkucuksen and Y. Altunbasak, “Edge strength filter based color filter array interpolation,” IEEE Trans. Image Process., 21 (1), 393 –397 (2012). https://doi.org/10.1109/TIP.2011.2155073 IIPRE4 1057-7149 Google Scholar

8. 

L. Chang and Y.-P. Tam, “Effective use of spatial and spectral correlations for color filter array demosaicing,” IEEE Trans. Consum. Electron., 50 (1), 355 –365 (2004). https://doi.org/10.1109/TCE.2004.1277885 ITCEDA 0098-3063 Google Scholar

9. 

W.-J. Chen and P.-Y. Chang, “Effective demosaicking algorithm based on edge property for color filter arrays,” Digital Signal Process., 22 (1), 163 –169 (2012). https://doi.org/10.1016/j.dsp.2011.09.006 DSPREJ 1051-2004 Google Scholar

10. 

X. Chen, G. Jeon and J. Jeong, “Voting-based directional interpolation method and its application to still color image demosaicking,” IEEE Trans. Circuits Syst. Video Technol., 24 (2), 255 –262 (2014). https://doi.org/10.1109/TCSVT.2013.2255421 Google Scholar

11. 

C.-Y. Su and W.-C. Kao, “Effective demosaicing using subband correlation,” IEEE Trans. Consum. Electron., 55 (1), 199 –204 (2009). https://doi.org/10.1109/TCE.2009.4814435 ITCEDA 0098-3063 Google Scholar

12. 

D. Kiku et al., “Minimized-Laplacian residual interpolation for color image demosaicking,” Proc. SPIE, 9023 90230L (2014). https://doi.org/10.1117/12.2038425 PSISDG 0277-786X Google Scholar

13. 

F. Godtliebsen, E. Spjotvoll and J. S. Marron, “A nonlinear Gaussian filter applied to images with discontinuities,” J. Nonparametric Stat., 8 21 –43 (1997). https://doi.org/10.1080/10485259708832713 NOSTEK 1048-5252 Google Scholar

14. 

I. Pekkucuksen and Y. Altunbasak, “Multiscale gradients-based color filter array interpolation,” IEEE Trans. Image Process., 22 (1), 157 –165 (2013). https://doi.org/10.1109/TIP.2012.2210726 IIPRE4 1057-7149 Google Scholar

15. 

Kodak lossless true color image suite,” (2018) http://r0k.us/graphics/kodak/ April ). 2018). Google Scholar

16. 

Image engineering,” (2018) http://www.image-engineering.de/ April ). 2018). Google Scholar

17. 

X. Zhang and B. A. Wandell, “A spatial extension of CIELAB for digital color image reproduction,” J. Soc. Inf. Disp., 5 (1), 61 –63 (1997). https://doi.org/10.1889/1.1985127 JSIDE8 0734-1768 Google Scholar

18. 

L. Zhang et al., “FSIM: a feature similarity index for image quality assessment,” IEEE Trans. Image Process., 20 (8), 2378 –2386 (2011). https://doi.org/10.1109/TIP.2011.2109730 IIPRE4 1057-7149 Google Scholar

Biography

Jin Wang received her BS degree in mathematics and applied mathematics from Zhejiang University, Hangzhou, China, and her MS and PhD degrees in electronic communications engineering from Hanyang University, Seoul, Korea, in 2007 and 2013, respectively. Currently, she is an associate professor at Xidian University in Xi’an, Shaanxi, China. She has been a researcher in the Department of Electronics and Computer Engineering, Hanyang University and the Department of Embedded Systems Engineering, Incheon University.

Jiaji Wu received his BS degree in electrical engineering from Xidian University, Xi’an, China, in 1996, his MS degree from National Time Service Center, the Chinese Academy of Sciences in 2002, and his PhD in electrical engineering from Xidian University in 2005. Currently, he is a professor at Xidian University, Xi’an China. His current research interests include image processing, still image coding, hyperspectral/multispectral image compression, communication, and high performance computing.

Zhensen Wu received his BSc degree in applied physics from Xi’an Jiaotong University, Xi’an, China, in 1969, and his MSc degree in space physics from Wuhan University, Wuhan, China, in 1981. Currently, he is a professor at Xidian University, Xi’an, China. From 1995 to 2001, he was invited multiple times as a visiting professor to Rouen University, France, for implementing joint study of two projects supported by the Sino-France Program for Advanced Research.

Marco Anisetti is an assistant professor at the Università degli Studi di Milano. His research interests are in the area of computational intelligence and its application to the design of complex systems and services. He investigates/has investigated the adoption of advanced techniques based on Kalman filters, rough sets, and fuzzy set theories for: (i) low-level functionalities like display management, advanced image processing and mobile geolocation, and (ii) high-level applications in the areas of humanized computing, human machine interaction and ambient intelligence.

Gwanggil Jeon received his BS, MS, and PhD degrees from Hanyang University, in 2003, 2005, and 2008, respectively. From 2009 to 2011, he was a postdoctoral fellow at University of Ottawa, and from 2011 to 2012, he was an assistant professor at Niigata University. Currently, he is a professor at Incheon National University. His research interests fall under the umbrella of image processing, particularly image compression, motion estimation, demosaicking, and image enhancement as well as computational intelligence such as fuzzy and rough sets theories.

© 2018 Society of Photo-Optical Instrumentation Engineers (SPIE) 0091-3286/2018/$25.00 © 2018 SPIE
Jin Wang, Jiaji Wu, Zhensen Wu, Marco Anisetti, and Gwanggil Jeon "Bayesian method application for color demosaicking," Optical Engineering 57(5), 053102 (9 May 2018). https://doi.org/10.1117/1.OE.57.5.053102
Received: 1 November 2017; Accepted: 4 April 2018; Published: 9 May 2018
Lens.org Logo
CITATIONS
Cited by 19 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Optical engineering

Image quality

Visualization

Statistical analysis

Digital filtering

Reconstruction algorithms

Color difference

RELATED CONTENT

Evaluating super resolution algorithms
Proceedings of SPIE (January 24 2011)

Back to Top