Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

BM3D-based denoising method for color polarization filter array

Open Access Open Access

Abstract

Color split-focal plane polarization imaging systems are composed of image sensors with a color polarization filter array (CPFA). The noise generated during image acquisition leads to incorrect estimation of the color polarization information. Therefore, it is necessary to denoise CPFA image data. In this study, we propose a CPFA block-matching and 3D filtering (CPFA-BM3D) algorithm for CPFA image data. The algorithm makes full use of the correlation between different polarization channels and different color channels, restricts the grouping of similar 2D image blocks to form 3D blocks, and attenuates Gaussian noise in the transform domain. We evaluate the denoising performance of the proposed algorithm using simulated and real CPFA images. Experimental results show that the proposed method significantly suppresses noise while preserving the image details and polarization information. Its peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) indicators are superior to those of the other existing methods. The mean values of the PSNR and SSIM of the degree of linear polarization (DoLP) color images calculated through CPFA image interpolation can be increased to 200% and 400%, respectively, by denoising with the proposed method.

© 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

1. Introduction

The main physical properties of light are the intensity, wavelength, coherence, and polarization. Among these, the human vision system can easily perceive intensity and wavelength information but cannot directly distinguish polarization. As physical properties such as those related to the object shape and surface roughness can be inferred from polarization information [1], it has been extensively applied in the fields of material classification [2], image-to-fog [3], 3D shape reconstruction [4], and biomedical imaging [5]. Polarization imaging systems can be classified as division of time (DoT) [6], division of amplitude, division of aperture, and division of focal plane (DoFP) systems [7]. The DoFP polarization imaging system has obvious advantages in terms of the system volume and imaging speed, which is the focus of research in the field of polarization imaging. Recently, newly released split-focal plane color polarizing cameras such as the FLIR BFS-U3-15S5PC and PHX050S cameras have produced more detailed and richer red-green-blue color polarization images than grayscale polarizing images. These cameras are equipped with an RGGB Bayer filter and directional polarization filter, which can form a 16-pixel computing unit as a color polarization filter array (CPFA), as shown in Fig. 1.

 figure: Fig. 1.

Fig. 1. Color focal plane polarized camera: (a) SONY IMX250MYRCMOS sensor, and RGGB Bayer filter and directional polarization filter. (b) Color polarization filter array (CPFA) with a 16-pixel computing unit.

Download Full Size | PDF

In the imaging process of a color focal plane polarization camera, noise interference from the outside world and the device itself, which generally includes noise from many different sources, is inevitable. As the noise characteristics tend to follow a Gaussian distribution when the number of noise sources increases, and the Gaussian noise model is easy to use in the design and analysis of the denoising algorithm [8], this study mainly examines the effect of Gaussian noise. The presence of noise degrades the image quality and produces spurious polarization information during interpolation reconstruction [9,10]. The color polarization interpolation algorithm [1113] is crucial for obtaining full-resolution color polarization images. It mainly restores all the 12-channel $({({R,G,B} )\ast ({{0^ \circ },{{45}^ \circ },{{90}^ \circ },{{135}^ \circ }} )} )$ images from a CPFA image dataset. However, these interpolation algorithms are all applied on the premise of noiseless CPFA images. Although denoising can be performed using denoising methods for color images after interpolation, it is more difficult to remove noise because interpolation changes the noise characteristics. A similar problem exists in grayscale DoFP images obtained using a directional polarization filter. Existing gray DoFP image denoising algorithms include those based on principal component analysis (PCA) [14], polarimetric denoising residual dense networks (PDRDN) [15], and block-matching and 3-D filtering (DoFP-BM3D) [16]. The PCA denoising algorithm utilizes reduction and linear least mean square errors to denoise in the PCA domain. The algorithm sacrifices the image details and edge information and is unsuitable for considerable Gaussian noise. The PDRDN denoising algorithm is suitable for image restoration in complex and high-noise environments. The DoFP-BM3D denoising algorithm is a universal DoFP image denoising method that can suppress most of the noise in an image, although certain image edge details are sacrificed. All these algorithms, however, are based on the DoFP image sensor. The CPFA image sensor adds color channels and changes the original image array format, which means these algorithms cannot be directly applied to CPFA image denoising. Qiu et al. first considered the influence of noise in the reconstruction process of color polarization images [17] and proposed an algorithm for solving the Stokes vector of color polarization images by solving an inverse problem within the alternating direction method of multipliers framework. However, as the main purpose of the algorithm is to reconstruct the Stokes vector, its denoising ability is limited. Therefore, it is only suitable for CPFA image data under low-intensity noise. In this study, we first prove the correlation between the newly-added color channel and the polarization direction channel, and we then propose a CPFA image denoising algorithm, CPFA-BM3D, based on block-matching and 3-D filtering (BM3D) [18]. The proposed algorithm makes full use of the correlation between the four polarization directions and three-color channels. As shown in Fig. 1(b), based on the super pixel, it utilizes the nonlocal self-similarity of the CPFA image and the correlation of the polarization and color information for finding similar blocks to form 3D blocks and subsequently performs 3D transformation and collaborative filtering. The performance of the denoising method is evaluated using the degree of linear polarization (DoLP) images reconstructed from simulated and real CPFA images. The obtained results establish that the proposed algorithm significantly improves the visual effect, peak signal-to-noise ratio (PSNR), and structural similarity (SSIM) [19].

The rest of the paper is organized as follows. In Section 2, we first prove the existence of correlation between the 12 channels in CPFA image data. We then introduce the block matching and 3D filtering (CPFA-BM3D) algorithm for CPFA image data and the algorithm parameter settings. In Section 3, we report our experimental results and conduct a comparison against competing denoising techniques in the literature. In Section 4, we make some concluding remarks about this work and future research.

2. Method

2.1 Correlation between color polarization channels

Interchannel correlation implies strong correlation between the high-frequency (texture or edge) components of different channels. The CPFA sensor contains not only different color channels (R, G, and B channels) but also different polarization channels (${0^ \circ },\,{45^ \circ },\,{90^ \circ }$, and ${135^ \circ }$ channels), which are arranged according to certain rules. By fully utilizing the correlation between these channels, excellent results can be achieved when the BM3D algorithm is applied in CPFA image data denoising.

We used the color polarization dataset in [17] as the test set. This dataset contains 40 carefully calibrated polarization scenes, each containing four RGB color images in the polarization directions of ${0^ \circ },\,{45^ \circ },\,{90^ \circ }$, and, respectively (detailed in Section 3.1.1). We selected 12 channels $({({R,G,B} )\ast ({{0^ \circ },{{45}^ \circ },{{90}^ \circ },{{135}^ \circ }} )} )$ in the center region of the image, which excluded the 15 outermost pixels, to avoid the boundary effect induced by the registration step used on the raw images. Subsequently, the Pearson correlation coefficient (PCC) [20] was used to evaluate the correlation between channels, and the 40 scenarios in the dataset were averaged, as formulated by Eq. (1):

$$PCC[{{C^u},{C^v}} ]= \frac{{\sum\nolimits_p {({({C_p^u - {\mu^u}} )({C_p^v - {\mu^v}} )} )} }}{{\sqrt {\sum\nolimits_p {{{({C_p^u - {\mu^u}} )}^2}} } \sqrt {\sum\nolimits_p {{{({C_p^v - {\mu^v}} )}^2}} } }}. $$
where ${C^u}$ and ${C^v}$ refer to two different channels, $({u,v} )\in {\{{1, \ldots ,12} \}^2}$.

Figure 2 shows the PCC results among 12 channels of the color polarization image. In the case of the R channel, the correlation between the four channels with different polarization (${R_{{0^ \circ }}},\,{R_{{{45}^ \circ }}},{R_{{{90}^ \circ }}},\,{R_{{{135}^ \circ }}}$) is higher than those between these four channels and the other eight channels $({({G,B} )\ast ({{0^ \circ },{{45}^ \circ },{{90}^ \circ },{{135}^ \circ }} )} )$. In the case of the B or G color channel, considering the ${G_{{0^ \circ }}}$ channel as an example $PCC({G_{{0^ \circ }}},{B_{{0^ \circ }}})\,>\,PCC({G_{{0^ \circ }}},{G_{{{45}^ \circ }}})\,>\,PCC({G_{{0^ \circ }}},{B_{{{45}^ \circ }}})\,>\,PCC({G_{{0^ \circ }}},{G_{{{135}^ \circ }}})\,> PCC({G_{{0^ \circ }}},{B_{{{135}^ \circ }}})\,>\,PCC({G_{{0^ \circ }}},{G_{{{90}^ \circ }}})\,>\,PCC({G_{{0^ \circ }}},{B_{{{90}^ \circ }}})$ shows that the correlation is maximum when the polarization channels are the same ($PCC({G_{{0^ \circ }}},{B_{{0^ \circ }}})$ is maximum). When the polarization channels differ, the correlation between channels of the same color is greater than that between channels of different color (e.g., $PCC({G_{{0^ \circ }}},{G_{{{45}^ \circ }}}) > PCC({G_{{0^ \circ }}},{B_{{{45}^ \circ }}})$). In general, for the 12 channels, the least PCC is $PCC({R_{{0^ \circ }}},{G_{{{90}^ \circ }}})$=0.9738, and the average PCC is 0.9916 (greater than 0.8 is correlated), indicating that the 12 channels have high correlation, and the correlation between channels can be fully utilized in CPFA image denoising.

 figure: Fig. 2.

Fig. 2. Correlation between 12 channels in the color polarization image.

Download Full Size | PDF

2.2 CPFA - BM3D algorithm

The CPFA image denoising algorithm based on BM3D (CPFA-BM3D) is a 3D transform domain filtering algorithm for denoising, utilizing the similarity between image blocks and the correlation between channels. It includes two main steps: basic estimation and final estimation. Each step includes image block grouping, 3D collaborative filtering, and aggregation, where 3D collaborative filtering uses hard-threshold filtering and Wiener filtering. The denoising process is illustrated in Fig. 3.

 figure: Fig. 3.

Fig. 3. CPFA-BM3D denoising algorithm flow chart.

Download Full Size | PDF

2.2.1 Basic estimation

  • (1) Image block grouping

We first divide the noisy CPFA image into reference blocks ${Z_{XR}}$ sized $N_1^{ht} \times N_1^{ht}$ in suitable steps. In natural images, there is a correlation between pixels of each reference block. The presence of strong correlation between the color polarization channels has been proven, so a certain correlation between the pixels of each reference block of the CPFA image is also expected. To take full advantage of the polarization information and the correlation between different channels, each reference block includes pixels of at least four polarization channels and three color channels $({({R,G,B} )\ast ({{0^ \circ },{{45}^ \circ },{{90}^ \circ },{{135}^ \circ }} )} )$, as shown in Fig. 1(b).

Further, we search for the number of matching blocks in search range ${n^{ht}} \times {n^{ht}}$.When the distance between the matching block and reference block is less than the fixed threshold $\tau _{match}^{ht}$, the matching block is considered as a similar block ${Z_{Xi}}$, and the set of ${Z_{Xi}}$ is $S_{XR}^{ht}$. The ${l^2}$ norm is used for calculating the similarity, as depicted in Eq. (2):

$$d({{Z_{Xi}},{Z_{XR}}} )= \frac{{|{|{{Z_{XR}} - {Z_{Xi}}} |} |_2^2}}{{{{({N_1^{ht}} )}^2}}}. $$
where $d({{Z_{Xi}},{Z_{XR}}} )$ is the ${l^2}$ distance between ${Z_{XR}}$ and ${Z_{Xi}},\,{({N_1^{ht}} )^2}$, is the size of the reference block in the basic estimation step, and ${|{|\cdot |} |_2}$ is the ${l^2}$ norm.

When blocks with different colors and different polarization orientation configurations are combined and thresholded (for example, as shown on the left of Fig. 4(a), when blocks on the upper-left corner of the ${R_{{{90}^ \circ }}}$ sample are combined with blocks on the upper-left corner of the ${G_{{0^ \circ }}}$ sample or of the ${B_{{{90}^ \circ }}}$ sample), severe checkerboard artifacts are observed in regions with low chromatic aberration and less polarization information in the denoised DoLP color image, as shown on the right of Fig. 4(a). This is because when finding similar blocks, two image blocks of different arrays with few differences in the color and polarization information have similar pixel values, resulting in similar values of the ${l^2}$ distance. These are judged as similar blocks, causing a deviation in the reference block estimate. To solve this problem, we limit the grouping to configuration blocks with the same color and polarization direction (when the block on the upper-left corner of the ${R_{{{90}^ \circ }}}$ sample is combined with that on the upper-left corner of the ${R_{{{90}^ \circ }}}$ sample), as shown on the left of Fig. 4(b). From the image on the right of Fig. 4(b), when the array configuration of similar blocks is considered to correspond to the color as well as polarization channels, the denoised DoLP color images are of better quality and retain more image details and polarization information.

 figure: Fig. 4.

Fig. 4. (a) (left) Groups of blocks with different colors and different polarization channel configurations and (right) denoised DoLP images ($\sigma \textrm{ = }20$). (b) (left) Groups of blocks with the same color and the same polarization channel configurations and (right) denoised DoLP images ($\sigma \textrm{ = }20$).

Download Full Size | PDF

All similar blocks in $S_{XR}^{ht}$ are sorted in descending order of similarity, and the top ${M^{ht}}$ similar blocks are used to build the similarity group ${Z_{S_{XR}^{ht}}}$.

  • (2) Collaborative hard-thresholding

Collaborative filtering in basis estimation is noise attenuation by hard thresholding in the 3D transform domain. The applied 3D linear transformation is called $\tau _{3D}^{ht}$, which is a portable 3D transformation. Wavelet transform or DCT 2D transform is performed first, followed by 1D Hadamard transform. The formula for collaborative hard-threshold filtering is as follows:

$${\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over Y} _{S_{XR}^{ht}}} = \tau {_{3D}^{ht- 1}}({\gamma ({\tau_{3D}^{ht}({{Z_{S_{XR}^{ht}}}} )} )} ). $$
where $\tau {_{3D}^{ht - 1}}$ represents the inverse transformation of $\tau _{3D}^{ht}$ and $\gamma ({\cdot} )$ represents the synergistic hard-threshold filter factor.

When collaborative filtering is applied to ${Z_{S_{XR}^{ht}}}$ composed of blocks with the same color and channel configuration, 3D transformation and noise removal is realized more effectively. This is because ${Z_{S_{XR}^{ht}}}$ has two unique advantages:

  • a) There is a certain correlation between pixels within each similar block.
  • b) There is a certain correlation among the corresponding pixels of all the similar blocks in ${Z_{S_{XR}^{ht}}}$, which we improved by setting a similar block array configuration with the same channel.

Collaborative filtering in basic estimation and final estimation utilizes these two types of correlation to generate a sparse representation of more real signals in similar block groups. Such sparse representation is highly effective in attenuating noise, retaining the signal characteristics, and realizing a more accurate filtering effect. Here, it is noted that filtering is applied to a 3D group containing all the polarization information and color information samples. Therefore, our method takes advantage of the correlation between all the channels.

Because multiple estimates can be found for the same pixel in each similar group, we assign smaller weights to noisy block estimates, and the basic estimate of each reference block is $\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over Y} _{{x_m}}^{ht,XR}$.

  • (3) Aggregation

Finally, the basic estimation image is obtained by clustering. The aggregation process can be expressed as follows:

$${\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over y} ^{basic}}(x )= \frac{{\sum\nolimits_{XR \in X} {\sum\nolimits_{{x_m} \in S_{XR}^{ht}} {w_{XR}^{ht}\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over Y} _{{x_m}}^{ht,XR}(x )} } }}{{\sum\nolimits_{XR \in X} {\sum\nolimits_{{x_m} \in S_{XR}^{ht}} {w_{XR}^{ht}{x_{{x_m}}}(x )} } }},\forall x \in X. $$
where ${x_{{x_m}}}({\cdot} )$ is the characteristic function of similar blocks. Weight $w_{XR}^{ht}$ is given by
$$w_{XR}^{ht} = \left\{ {\begin{array}{c} {\frac{1}{{{\sigma^2}N_{har}^{XR}}},N_{har}^{XR} \ge 1}\\ {1,otherwise} \end{array}} \right.$$
where $N_{har}^{XR}$ is the number of nonzero elements in the sparse representation of hard- thresholding.

2.2.2 Final estimation

In the final estimation, the original noisy image is Wiener-filtered, mainly based on the basic- estimate image. This step is primarily applied to restore additional image details and improve the denoising performance of the algorithm. The parameters in the final estimation are denoted by superscript “wie”. Similar blocks are grouped in the estimated image and the noise image. The specific grouping method is the same as that used in basic estimation. We assume a set of two similar blocks: $\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over Y} _{S_{XR}^{wie}}^{basic}$ and $S_{XR}^{wie}$.

The contraction coefficient of the Wiener filter is defined by the energy spectrum of the 3D transformation coefficient of the basic estimation group:

$${W_{S_{XR}^{wie}}} = \frac{{{{\left|{\tau_{3D}^{wie}\left( {\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over Y}_{S_{XR}^{wei}}^{basic}} \right)} \right|}^2}}}{{{{\left|{\tau_{3D}^{wie}\left( {\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over Y}_{S_{XR}^{wei}}^{basic}} \right)} \right|}^2} + {\sigma ^2}}}. $$

All the similar blocks in $S_{XR}^{wie}$ are sorted in the order of similarity, and the 3D group ${Z_{S_{XR}^{wie}}}$ is constructed with the first ${M^{wie}}$ similar blocks in the sort. Noise data are processed using 3D transformation and Wiener coefficient ${W_{S_{XR}^{wie}}}$, followed by 3D inverse transformation to obtain the estimated value of the set of similar image blocks:

$$\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over Y} _{S_{XR}^{wie}}^{wie} = \tau {_{3D}^{wie - 1}}({{W_{S_{XR}^{wie}}}\tau_{3D}^{wie}({{Z_{S_{XR}^{wie}}}} )} ). $$

The final estimate is obtained as the weighted average of different estimates of the same reference block:

$${\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over y} ^{final}}(x )= \frac{{\sum\nolimits_{XR \in X} {\sum\nolimits_{{x_m} \in S_{XR}^{ht}} {w_{XR}^{wie}\mathord{\buildrel{\lower3pt\hbox{$\scriptscriptstyle\frown$}} \over Y} _{{x_m}}^{wie,XR}(x )} } }}{{\sum\nolimits_{XR \in X} {\sum\nolimits_{{x_m} \in S_{XR}^{wie}} {w_{XR}^{wie}{x_{{x_m}}}(x )} } }},\forall x \in X. $$
where $ is the pixel estimation of image block $ at position X in the final estimated- image similarity block group and $w_{XR}^{wie}$ is the weight of each similar block given by:
$$w_{XR}^{wie} = {\sigma ^{ - 2}}|{|{{W_{S_{XR}^{wie}}}} |} |_2^{ - 2}. $$

2.3 Parameter setting

Parameter selection generally affects the performance of an algorithm. Because the correlation of the 12 channels in the CPFA image imposes certain restrictions on the parameter setting, we studied the parameter setting of the CPFA-BM3D algorithm comprehensively considering the algorithm performance, running time, correlation, and so on. The parameters in the basic estimation step are represented by superscript “ht”, whereas those in the final estimation step are represented by superscript “wie” as previously mentioned. The parameter set for the basic and final estimation, respectively, are:

  • $N_1^{ht}$ and $N_1^{wie}$: size of the reference block and similar block;
  • ${M^{ht}}$ and ${M^{wie}}$: maximum number of similar blocks;
  • ${p^{ht}}$ and ${p^{wie}}$: selected step size of the reference image;
  • ${n^{ht}}$ and ${n^{wie}}$: search window size;
  • $\tau _{match}^{ht}$ and $\tau _{match}^{wie}$: maximum threshold of the distance between two similar blocks.

After numerous experiments, we found that several parameters, namely, ${n^{ht}},\,{n^{wie}},\,\tau _{match}^{ht}$, and $\tau _{match}^{wie}$, had negligible effect on the result. Therefore, the following parameter values were fixed and used throughout the study: ${n^{ht}} = 39,\,{n^{wie}} = 39,\,\tau _{match}^{ht} = 2500$, and $\tau _{match}^{wie} = 400$.

We tested all 40 scene images in the test set. To quantitatively describe the denoising results, the PSNR and SSIM (detailed in Section 3.1.2) were used as the two evaluation indices. We compared five groups of images, including the Stokes vectors (S0, S1, S2), DoLP, and the angle of polarization (AoP) obtained before and after CPFA image denoising. The results showed that the parameter setting results for the five groups of compared images in the 40 scenes were similar. For simplicity, we consider the result of the DoLP image of a plant scene as an example to illustrate the parameter setting. In addition, when other parameter values are not given, we use the optimal solution for those parameters.

2.3.1 Influence of $N_1^{ht}$ and $N_1^{wie}$

To maintain the integrity of the color polarization array in the reference block, and considering the correlation between the color polarization channels, the selection of the reference block has certain requirements, owing to the unique structure of CPFA image data. When $N_1^{ht} = 2$ or $N_1^{wie} = 2$, the denoising effect is very poor because all the channels are not selected in the reference block, and the collaborative filtering effect is poor because the reference block is very small. Therefore, $N_1^{ht},N_1^{wie} \ge 4$ were set. For a low noise standard deviation, the reference block must be relatively small to preserve image details well. However, for a larger noise standard deviation, the denoising effect is better when the reference block is larger as most of the image details are lost to noise.

As shown in Table 1, when $\sigma = 30$ and the other denoising algorithms have the same parameters, the reference block size can result in a difference of 1.625 dB in the PSNR and 0.053 dB in the SSIM. Simultaneously, the size of the reference block affects the execution speed of the algorithm. The smaller the reference block, the faster is the algorithm. For $\sigma \le 20$ and $N_1^{ht} = 8,\,N_1^{wie} = 8$ and $N_1^{wie} = 12$ are very close. Therefore, we selected $N_1^{wie} = 8$ for its faster execution time.

Tables Icon

Table 1. Influence of $N_1^{ht}$ and $N_1^{wie}$

2.3.2 Influence of the ${M^{ht}}$ and ${M^{wie}}$

As shown in Table 2, when the noise is relatively low, the maximum number of similar blocks has negligible influence on the denoising result, whereas it has a significant influence when the noise is relatively high. The denoising effect is best when ${M^{ht}}\textrm{ = }32$ and ${M^{wie}}\textrm{ = }64$.

Tables Icon

Table 2. Influence of ${M^{ht}}$ and ${M^{wie}}$.

2.3.3 Influence of ${p^{ht}}$ and ${p^{wie}}$

To increase the processing speed, a circular selection of reference image blocks is performed with step size p (integer) in rows and columns. For example, with $p = 1$, the processing is theoretically nine times faster than with $p = 3$. In addition, because the final estimation is based on the basic estimate that assumes no noise or at least low noise, the step-size of the final estimate can be larger than that of the basic. From Table 3, ${p^{ht}} = 3$ and ${p^{wie}} = 5$ gives the best result.

Tables Icon

Table 3. Influence of ${p^{ht}}$ and ${p^{wie}}$.

To increase the processing speed, a circular selection of reference image blocks is performed with step size p (integer) in rows and columns. For example, with $p = 1$, the processing is theoretically nine times faster than with $p = 3$. In addition, because the final estimation is based on the basic estimate that assumes no noise or at least low noise, the step-size of the final estimate can be larger than that of the basic. From Table 3, ${p^{ht}} = 3$ and ${p^{wie}} = 5$ gives the best result.

3. Experimental results and discussion

3.1 Experimental preparation

3.1.1 Simulated and real CPFA images

In this study, the color polarization dataset in Ref. [17], containing 40 carefully calibrated polarization scenes obtained using a DoT polarization imaging system, was used as the test dataset. The DoT polarization imaging system changes the wavelength and polarization angle of light by rotating the spectral filter and polarization filter multiple times for combining different spectral and polarization filters to obtain the color polarization images. This system can directly obtain each pixel including 12 values $({({R,G,B} )\ast ({{0^ \circ },{{45}^ \circ },{{90}^ \circ },{{135}^ \circ }} )} )$ full-resolution image, mainly used for verification of the principle of polarization imaging. For each scene, there are four groups of raw images at polarizations of ${0^ \circ },\,{45^ \circ },\,{90^ \circ }$, and ${135^ \circ }$ with the same exposure time. Each group consists of 100 images acquired continuously under the same situation, averaged to suppress noise. The resolution of each image is 1024 ${\times} $1024 pixel and the depth are 8 bits. To construct the simulated CPFA images, the images of 12 channels $({({R,G,B} )\ast ({{0^ \circ },{{45}^ \circ },{{90}^ \circ },{{135}^ \circ }} )} )$ were down-sampled, and simulated noiseless CPFA images were generated according to the array mode shown in Fig. 1(b), whose sizes were equal to those of the four color-images. Further, Gaussian white noise was added to the simulated noiseless CPFA images, which were then denoised through means filtering, median filtering, Wiener filtering, wavelet threshold filtering [21], the BM3D algorithm, the DoFP-BM3D algorithm, the algorithm proposed by Qiu et al. (hereafter called Qiu’s method), and our proposed CPFA-BM3D algorithm. The BM3D algorithm extracts the images of the 12 channels from the CPFA image data, denoises the images of each channel separately, and then synthesize the denoised images into CPFA image data. The DoFP-BM3D algorithm extracts the images of each color channel in the CPFA image to synthesize four (R, G, G, B) DoFP images, denoising them separately, and then synthesizes the denoised subimages into CPFA image data. The other denoising algorithms are directly applied to the CPFA image data. Among them, mean filtering, median filtering, and Wiener filtering adopt a 3*3 window. Wavelet threshold filtering, the BM3D algorithm, and Qiu’s method employ algorithms with source code provided in references [21], [18], and [17], respectively. The DoFP-BM3D algorithm adopts the parameters in Ref. [16]. Finally, the denoising performance was evaluated based on the color PSNR (average value of the RGB channels), SSIM, and visual comparison.

In addition, we used a Lucid Vision Lab PHX050S-Q color polarization camera to record real CPFA images. The $2048 \times 2448$ pixel SONY IMX250MYR CMOS color polarization sensor was used, where each pixel was $3.45\mu m \times 3.45\mu m$, and the depth was 8-bit.

3.1.2 Evaluation indices

The PSNR is defined as follows:

$$PSNR = 10 \cdot {\log _{10}}\left( {\frac{{MA{X^2}}}{{MSE}}} \right)$$
where MAX represents the maximum value of the image point pixel, and MSE is defined as
$$MSE = \frac{1}{{mn}}\sum\limits_{i = 0}^{m - 1} {\sum\limits_{j = 0}^{n - 1} {{{[{I({i,j} )- K({i,j} )} ]}^2}} }. $$

Here, $m \times n$ is the size of the image, and $I({i,j} )$ and $K({i,j} )$ represent the simulated real image and the denoised image, respectively.

The SSIM measures the image similarity with respect to three aspects: brightness, contrast, and structure. It is defined as follows:

$$SSIM({x,y} )= \frac{{({2{\mu_x}{\mu_y} + {c_1}} )({2{\sigma_{xy}} + {c_2}} )}}{{({\mu_x^2 + \mu_y^2 + {c_1}} )({\mu_x^2 + \mu_y^2 + {c_2}} )}}. $$
where ${\mu _x}$ and ${\mu _y}$ are the average of x and y, respectively, $\sigma _x^2$ and $\sigma _y^2$ are the variance of x and y, respectively, ${\sigma _{xy}}$ is the covariance of x and y, and ${c_1}$ and ${c_2}$ are constants.

3.2 Experimental results

In this section, the simulated images are compared in terms of the PSNR, SSIM, and visually, whereas the real images are only compared visually. Considering the denoising effect and computational complexity, the algorithm parameters were set as follows: $N_1^{ht} = 8$, $N_1^{wie} = 8,\,{M^{ht}} = 32,\,{M^{wie}} = 64,\,{p^{ht}} = 3,\,{p^{wie}} = 5,\,{n^{ht}} = 39,\,{n^{wie}} = 39,\,\tau _{match}^{ht} = 2500$, and $\tau _{match}^{wie} = 400$.

3.2.1 Results of the simulated CPFA images

We added Gaussian white noise ($\sigma \textrm{ = 20}$) to the simulated CPFA images and denoised them using the different denoising methods mentioned in Section 3.1.1. The edge-aware residual interpolation [12] color polarization interpolation algorithm was used to interpolate the denoised simulated CPFA images to generate the color images linearly polarized in four directions. The sizes of these linearly polarized color images were equal to those of the simulated CPFA images. The DoLP color image calculated from the original color polarization images with four different polarization directions in the dataset was used as the reference image. The DoLP color image calculated from the four interpolated images was compared with the reference image. The denoising results of the DoLP color images of the “plant” scene in the test set are shown in Fig. 5.

 figure: Fig. 5.

Fig. 5. Simulated CPFA image denoising results: (a) original images, (b) nonuniformity corrected images, (c) average images, (d) median filtering, (e) Wiener filtering, (f) wavelet threshold filtering, (g) Qiu’s method, (h) BM3D, (i) DoFP-BM3D, and (j) our method.

Download Full Size | PDF

Figure 5(a1) displays the DoLP reference image, and Fig. 5(b1) displays the DoLP image obtained by interpolating the simulated CPFA image with Gaussian noise (σ=20). Figures 5(c1)–(j1), respectively, show the results of mean filtering, median filtering, Wiener filtering, wavelet threshold filtering, Qiu’s method, BM3D algorithm, DoFP-BM3D algorithm, and the DoLP image obtained by interpolation calculation after denoising the simulated CPFA image using the CPFA-BM3D algorithm. Figures 5(a2)–(j2) display the partial enlargement of the green box in the respective DoLP color image.

As shown in Figs. 5(c1)–(f1), after denoising the simulated CPFA image using mean filtering, median filtering, Wiener filtering, and wavelet threshold filtering, the noise in the DoLP color image is partly reduced compared to the DoLP color image with noise (Fig. 5(b1)). However, some of the areas in the image are distorted because the residual noise in the denoised image is amplified by the Stokes parameter calculation. In addition, in Figs. 5(c2)–(f2), the texture and detail information of the leaves and flower-pots are severely corrupted by noise, affecting the quality of the DoLP color images. Figures 5(g1)–(g2) indicate that when $\sigma $ is large ($\sigma \textrm{ = 20}$), the DoLP color image after denoising by Qiu’s method is filled with noise, and there is no effective polarization information in the image. From Figs. 5(h1)–(j1), the image quality of the DoLP color images denoised using the BM3D, DoFP-BM3D, and CPFA-BM3D algorithms is significantly improved, and the difference in polarization between the leaves and flowerpots can be observed. However, in Fig. 5(h2), the leaf-edge of the DoLP color image denoised by the BM3D algorithm has obvious distortion, and checkerboard artifacts can be observed on the flowerpot. As shown in Fig. 5(i2), although the DoFP-BM3D denoising algorithm reduces distortion of the leaf-edges, checkerboard artifacts are present in the leaf and flowerpot because only the polarization information is utilized in this method and color correlation is not applied. As shown in Fig. 5(j2), compared to the other algorithms, the denoised DoLP color image using our proposed CPFA-BM3D algorithm removes noise more thoroughly, preserves the edge details of the leaf and flowerpot in the enlarged region more completely, and reduces the error in the DoLP color image. Therefore, in terms of the visual effects, the denoising effect of our algorithm is significantly better than those of the other denoising algorithms.

To verify the denoising effect of different algorithms at different noise levels, Gaussian white noise with different standard deviations was added to the simulated CPFA image “plant”. The PSNR and SSIM of the DoLP color image after denoising with different algorithms are plotted in Fig. 6.

 figure: Fig. 6.

Fig. 6. PSNR (left) and SSIM (right) under different noise standard deviations.

Download Full Size | PDF

From Fig. 6, under small standard deviation ($\sigma < 10$), the difference in the PSNR and SSIM of the DoLP color image before and after denoising are negligible, and the denoising effect of each algorithm is almost the same. However, our denoising algorithm is always the best. When the standard deviation of noise increases, the PSNR and SSIM of the DoLP color image denoised by our CPFA-BM3D algorithm remain superior to those of the other algorithms. The advantage of our algorithm over the other algorithms becomes increasingly obvious with the increase in the standard deviation of noise. Moreover, the PSNR and SSIM of the image after denoising by our algorithm are always greater than 32 dB and 0.85, respectively, indicating that our algorithm is effective for different noise levels, has strong adaptability, and a more obvious suppression effect on excessive noise.

The polarization information of color-polarized images is generally embodied in S0, S1, S2, the DoLP, and AoP. Therefore, the performance of the denoising algorithm can be evaluated using these five types of images. We denoised all 40 simulated CPFA images in the test set ($\sigma \textrm{ = 20}$) and averaged the denoising results of the 40 sets of scenes. The PSNR and SSIM values for the five images are listed in Table 4.

Tables Icon

Table 4. Mean PSNR and mean SSIM of the S0, S1, S2, DoLP, and AoP denoising results on the test set ($\sigma \textrm{ = 20}$).

Comparing the values of the PSNR and SSIM in Table 4, it is obvious that our algorithm outperforms the other algorithms. For S1 and S2 images, our algorithm improves the PSNR and SSIM by approximately 4 dB and 0.2, respectively, compared to the other algorithms. Compared to the noisy images, the mean value of the PSNR of DoLP color images after denoising can be increased to 200%, whereas that of AoP color images after denoising can be increased to 160%. The mean SSIM of DoLP color images is increased to 400%. The above results show that our algorithm is highly effective in preserving the polarization information of color-polarized images.

3.2.2 Results of real CPFA images

For real CPFA images, correction is first performed for nonuniformity [22], after which the noise estimation algorithm [23] is used to estimate the noise of sub-images with different polarization directions and different color channels to obtain the standard deviation of the sub-image noise. Finally, the image is denoised and interpolated to obtain the computed DoLP color image.

Figure 7(a1) shows the DoLP color image obtained by interpolation calculation of a real CPFA image, and Fig. 7(b1) shows the result of nonuniformity correction. Figures 7(c1)–(j1) depict the DoLP color image after denoising a real CPFA image, and Figs. 7(a2)–(j2) show the partial enlargement of the green box in the respective DoLP color image. In Fig. 7(a1), there are severe inhomogeneities in the contours of the sculptures and buildings in the image, and there is considerable noise. As shown in Fig. 7(b1), the image quality after nonuniformity correction is significantly improved, but considerable noise is present, resulting in no significant difference in the polarization of the object. Figures 7(c1)–(f1) indicate that the image quality of the DoLP color image after denoising by mean filtering, median filtering, Wiener filtering, and wavelet threshold filtering, respectively, is improved. However, some image edge details are lost, resulting in excessive smoothness. Figures 7(h2) and 7(i2) show that the BM3D and DoFP-BM3D algorithms can remove part of the noise, but there is still some noise left in the image and there is severe loss of polarization information at the railings. Figures 7(g2) and 7(j2) demonstrate that our CPFA-BM3D algorithm and Qiu’s algorithm have better results overall than the other algorithms. However, our CPFA-BM3D algorithm is clearly better with respect to the detailed processing of the sculpture edges compared to Qiu’s algorithm. The polarization difference of the sculptures in the enlarged region is more obvious. The polarization information of the DoLP color images is retained, whereas the DoLP images processed by Qiu’s algorithm are smooth and lack details. Moreover, the denoising results of the simulated and real CPFA images indicate that Qiu’s algorithm is unsuitable for high-intensity noise. Therefore, our algorithm is more suitable for real CPFA-image denoising compared to the other algorithms.

 figure: Fig. 7.

Fig. 7. Real CPFA-image denoising results: (a) original images, (b) nonuniformity corrected images, (c) average images, (d) median filtering, (e) wiener filtering, (f) wavelet threshold filtering, (g) Qiu’s method, (h) BM3D, (i) DoFP-BM3D, and (j) our method.

Download Full Size | PDF

4. Conclusion

In this study, we proved the existence of correlation between 12 CPFA image data channels and proposed a BM3D-based denoising method for CPFA image data (CPFA-BM3D). This method makes full use of the correlation between different polarization channels and different color channels to limit the search for similar blocks and groups blocks corresponding to the same color channel and polarization channel to reduce chromatic aberration and checkerboard artifacts. Subsequently, 3D transformation and filtering were performed for similar block groups. Application of the correlation between different channels adds sparse representation of the real signals of similar block groups in the transform domain, further improving the filtering effect. Finally, the filtering results were aggregated, and the CPFA image was restored. The denoising performance of the proposed algorithm was evaluated using simulated and real CPFA images. The obtained results demonstrated that the proposed algorithm was superior to the existing algorithms in terms of the visual effect, PSNR, and SSIM, and that it could retain edge details and texture information. Moreover, the real CPFA image quality was obviously improved after processing with the proposed algorithm. The PSNR and SSIM of the DoLP color images obtained by interpolation of CPFA images processed by the proposed algorithm are increased to 200% and 400% of their original values, respectively. However, the detailed processing of this method is not ideal for non-Gaussian noise. Therefore, we intend to study more general CPFA image denoising methods and CPFA image interpolation denoising methods in the future.

Funding

Shanxi Province Colleges and Universities Science and Technology Innovation Project (2020L0048); the Natural Science Young Foundation of Shanxi Province (201901D211172); the State Council and the central government guide local funds of China (YDZX20201400001547).

Disclosures

The authors declare no conflicts of interest.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. B. Huang, T. Liu, H. Hu, and M. Yu, “Underwater image recovery considering polarization effects of objects,” Opt. Express 24(9), 9826–9838 (2016). [CrossRef]  

2. B. Matthieu, H. Hu, and G. François, “Contrast optimization in broadband passive polarimetric imaging,” Opt. Lett. 39(23), 6759–6762 (2014). [CrossRef]  

3. S. Fang, X. Xia, X. Huo, and C. Chen, “Image dehazing using polarization effects of objects and airtight,” Opt. Express 22(16), 19523–19537 (2014). [CrossRef]  

4. Z. Chen, X. Wang, and R. Liang, “Snapshot phase shift fringe projection 3D surface measurement,” Opt. Express 23(2), 667–673 (2015). [CrossRef]  

5. Z. Guan, F. Goodail, M. Yu, X. Li, Q. Han, Z. Cheng, H. Hu, and T. Liu, “Contrast optimization in broadband passive polarimetric imaging based on color camera,” Opt. Express 27(3), 2444–2454 (2019). [CrossRef]  

6. J. S. Tyo, D. L. Goldstein, D. B. Chenault, and J. A. Shaw, “Review of passive imaging polarimetry for remote sensing applications,” Appl. Opt. 45(22), 5453–5469 (2006). [CrossRef]  

7. R. Perkins and V. Gruev, “Signal-to-noise analysis of Stokes parameters in division of focal plane polarimeters,” Opt. Express 18(25), 25815–25824 (2010). [CrossRef]  

8. L. Zhang, R. Lukac, X. Wu, and D. Zhang, “PCA-based spatially adaptive denoising of CFA Images for single-sensor digital cameras,” IEEE Trans. Image Process. 18(4), 797–812 (2009). [CrossRef]  

9. A. Carnicer and B. Javidi, “Polarimetric 3D integral imaging in photon-starved conditions,” Opt. Express 23(5), 6408–6417 (2015). [CrossRef]  

10. X. Li, T. Liu, B. Huang, Z. Song, and H. Hu, “Optimal distribution of integration time for intensity measurements in Stokes polarimetry,” Opt. Express 23(21), 27690–27699 (2015). [CrossRef]  

11. S. Wen, Y. Zheng, F. Lu, and Q. Zhao, “Convolutional demosaicing network for joint chromatic and polarimetric imagery,” Opt. Lett. 44(22), 5646–5649 (2019). [CrossRef]  

12. M. Morimatsu, Y. Monno, M. Tanaka, and M. Okutomi, “Monochrome and color polarization demosaicking using edge-aware residual interpolation,” IEEE Sens. J. 21(23), 26985–26996 (2021). [CrossRef]  

13. S. Wen, Y. Zheng, and F. Lu, “A sparse representation based joint demosaicing method for single-chip polarized color sensor,” IEEE Trans. Image Process. 30, 4171–4182 (2021). [CrossRef]  

14. J. Zhang, H. Luo, R. Liang, W. Zhou, B. Hui, and Z. Chang, “PCA-based denoising method for division of focal plane polarimeters,” Opt. Express 25(3), 2391–2400 (2017). [CrossRef]  

15. X. Li, H. Li, Y. Lin, J. Guo, J. Yang, H. Yue, K. Li, C. Li, Z. Cheng, H. Hu, and T. Liu, “Learning-based denoising for polarimetric images,” Opt. Express 28(11), 16309–16321 (2020). [CrossRef]  

16. A. Abubakar, X. Zhao, S. Li, M. Takruri, E. Bastaki, and A. Bermak, “A block-matching and 3-D filtering algorithm for Gaussian noise in DoFP polarization images,” IEEE Sens. J. 18(18), 7429–7435 (2018). [CrossRef]  

17. S. Qiu, Q. Fu, C. Wang, and H. Wolfgang, “Linear polarization demosaicking for monochrome and colour polarization focal plane arrays,” Comput. Graph. Forum 40(6), 77–89 (2021). [CrossRef]  

18. K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, “Image denoising by sparse 3-D transform-domain collaborative filtering,” IEEE Trans. on Image Process. 16(8), 2080–2095 (2007). [CrossRef]  

19. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Trans. Image Process. 13(4), 600–612 (2004). [CrossRef]  

20. X. Li, B. Gunturk, and L. Zhang, “Image demosaicing: A systematic survey,” Proc. SPIE 6822, 68220F (2008). [CrossRef]  

21. D. L. Donoho and I. M. Johnstone, “Ideal spatial adaptation by wavelet shrinkage,” Biometrika 81(3), 425–455 (1994). [CrossRef]  

22. J. Zhang, H. Luo, B. Hui, and Z. Chang, “Non-uniformity correction for division of focal plane polarimeters with a calibration method,” Appl. Opt. 55(26), 7236–7240 (2016). [CrossRef]  

23. P. Stanislav, L. Zhang, and H. Jürgen, “Fast noise variance estimation by principal component analysis,” Proc. SPIE 8655, 86550N (2013). [CrossRef]  

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (7)

Fig. 1.
Fig. 1. Color focal plane polarized camera: (a) SONY IMX250MYRCMOS sensor, and RGGB Bayer filter and directional polarization filter. (b) Color polarization filter array (CPFA) with a 16-pixel computing unit.
Fig. 2.
Fig. 2. Correlation between 12 channels in the color polarization image.
Fig. 3.
Fig. 3. CPFA-BM3D denoising algorithm flow chart.
Fig. 4.
Fig. 4. (a) (left) Groups of blocks with different colors and different polarization channel configurations and (right) denoised DoLP images ($\sigma \textrm{ = }20$). (b) (left) Groups of blocks with the same color and the same polarization channel configurations and (right) denoised DoLP images ($\sigma \textrm{ = }20$).
Fig. 5.
Fig. 5. Simulated CPFA image denoising results: (a) original images, (b) nonuniformity corrected images, (c) average images, (d) median filtering, (e) Wiener filtering, (f) wavelet threshold filtering, (g) Qiu’s method, (h) BM3D, (i) DoFP-BM3D, and (j) our method.
Fig. 6.
Fig. 6. PSNR (left) and SSIM (right) under different noise standard deviations.
Fig. 7.
Fig. 7. Real CPFA-image denoising results: (a) original images, (b) nonuniformity corrected images, (c) average images, (d) median filtering, (e) wiener filtering, (f) wavelet threshold filtering, (g) Qiu’s method, (h) BM3D, (i) DoFP-BM3D, and (j) our method.

Tables (4)

Tables Icon

Table 1. Influence of N 1 h t and N 1 w i e

Tables Icon

Table 2. Influence of M h t and M w i e .

Tables Icon

Table 3. Influence of p h t and p w i e .

Tables Icon

Table 4. Mean PSNR and mean SSIM of the S0, S1, S2, DoLP, and AoP denoising results on the test set ( σ  = 20 ).

Equations (12)

Equations on this page are rendered with MathJax. Learn more.

P C C [ C u , C v ] = p ( ( C p u μ u ) ( C p v μ v ) ) p ( C p u μ u ) 2 p ( C p v μ v ) 2 .
d ( Z X i , Z X R ) = | | Z X R Z X i | | 2 2 ( N 1 h t ) 2 .
Y S X R h t = τ 3 D h t 1 ( γ ( τ 3 D h t ( Z S X R h t ) ) ) .
y b a s i c ( x ) = X R X x m S X R h t w X R h t Y x m h t , X R ( x ) X R X x m S X R h t w X R h t x x m ( x ) , x X .
w X R h t = { 1 σ 2 N h a r X R , N h a r X R 1 1 , o t h e r w i s e
W S X R w i e = | τ 3 D w i e ( Y S X R w e i b a s i c ) | 2 | τ 3 D w i e ( Y S X R w e i b a s i c ) | 2 + σ 2 .
Y S X R w i e w i e = τ 3 D w i e 1 ( W S X R w i e τ 3 D w i e ( Z S X R w i e ) ) .
y f i n a l ( x ) = X R X x m S X R h t w X R w i e Y x m w i e , X R ( x ) X R X x m S X R w i e w X R w i e x x m ( x ) , x X .
w X R w i e = σ 2 | | W S X R w i e | | 2 2 .
P S N R = 10 log 10 ( M A X 2 M S E )
M S E = 1 m n i = 0 m 1 j = 0 n 1 [ I ( i , j ) K ( i , j ) ] 2 .
S S I M ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( μ x 2 + μ y 2 + c 2 ) .
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.