Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

A Bayesian approach for polarimetric data reduction: the Mueller imaging case

Open Access Open Access

Abstract

In this paper, we extend to the Mueller imaging framework a formerly introduced Bayesian approach dealing with polarimetric data reduction and robust clustering of polarization encoded images in the piecewise constant case. The extension was made possible thanks to a suitable writing of the observation model in the Mueller context that relies on the system’s coherency matrix and Cholesky decomposition such that the admissibility constraints are easily captured. This generalization comes at the cost of nonlinearity with respect to the parameters that have to be estimated. This estimation-clustering problem is tackled in a Bayesian framework where a hierarchical stochastic model based on a Markov random field proposed by Potts is used. This fully unsupervised approach is extensively tested over synthetic data as well as real Mueller images.

©2008 Optical Society of America

1. Introduction

In a previous paper [1], we reported on a new method addressing polarimetric data reduction in the Stokes imaging framework. Our main contribution there concerned the use of a Bayesian approach ensuring the physical authenticity of polarization encoded images obtained from raw intensity acquisitions while providing a robust clustering of the imaged scene. We showed also that our approach preserved the Stokes channels from image noises and errors induced by the imaging system. The problem of the more complex Mueller imaging case remained pending. It is addressed in the present article.

The importance of Mueller imaging comes from the rich set of information it provides about the local nature of the target which may be of great use in studying number of physical phenomena that arise in many fields of science. Hence, an increasing interest is devoted to the development of new imaging systems that record spatially dependent patterns of polarized light diffusely scattered or transmitted by scenes or samples.

Mueller imaging is defined as the distributed measurement of Mueller matrices attached to each pixel. This is made possible by the use of effective polarization state analyzers (PSAs) that acquire the Stokes vectors corresponding to each pixel in the image while a source lights up the scene through a polarization state generator (PSG). Different configurations using different polarization modulation techniques are used in imaging polarimetry. Whatever the technique used, the principle remains the same: a set of raw intensity images have to be inverted to yield polarization encoded images. This procedure called “polarimetric data reduction” (PDR) is achieved through a linear observation model that relates the two sets of images via the system dependent PSA and PSG matrices.

It is worth noting that a proper choice of the PSA and the PSG matrices plays an important role in PDR [2]. It ensures a good conditioning that allows the inversion of the system and minimizes the contamination of polarimetric channels by image noise. However, if one uses the classical PDR techniques, even for well behaved imaging polarimeters [2], the image noise propagates into the estimated Mueller matrices and may prevent their physical admissibility. Indeed, a 4×4 real matrix must satisfy some physical constraints to be bona fide.

The contribution of the present work is twofold. Firstly, we propose an original parameterization of the forward model in order to account for the physical admissibility conditions. This leads to a nonlinear observation model. Secondly, we propose a fully unsupervised Bayesian framework (model and algorithm) addressing the estimation problem. We suppose that the sought object is piecewise constant, which corresponds to many situations of practical interest. The Markov random field model originally proposed by Potts will account for the hidden segmentation map holding the spatial prior knowledge. We will also assume the Mueller matrices to be class wise constant. Estimation will take place in the generalized maximum likelihood framework where the a posteriori probability law is optimized jointly with respect to the parameters and the hyper-parameters.

The article is organized as follows: the nonlinear parameterization of the forward problem and the stochastic model are presented in section 2. The estimation procedure is detailed in section 3. Simulation results and real data processing are presented and discussed in section 4. Conclusion and perspectives are given in section 5.

2. Problem statement

2.1. Observation model in the Mueller imaging polarimetry framework

A Mueller imaging polarimeter allows the indirect measurement of the Mueller matrix attached to each pixel in the image. This is achieved by illuminating an object by p independent polarization states where p≥4 through a PSG. Each of these states interacts with the object to yield an outgoing Stokes vector that is sensed through the PSA by q independent probing states where q≥4. For each pixel location (jy, jx) in the image, the p×q intensity measurements I are related to the pixel’s Mueller matrix M by the following linear equation:

I(jy,jx)=AM(jy,jx)W,

where A is a p×4 PSA matrix whose rows constitute the p probing states and W is the 4×q PSG matrix whose columns represent the q incident polarization states.

It can be shown [2] that Eq. (1) can be written in the following form

I¯(jy,jx)=(WtA)M¯(jy,jx)=P(η)M¯(jy,jx),

where the underline symbol stands for the operator which maps a p×q matrix into an pq×1 vector formed by stacking up its columns and ⊗ represents the Kronecker product. The known pq×16 matrix P, also called “Polarization Measurement Matrix” (PMM), depends on a parameter vector η which controls the combined actions of the PSG and PSA and relates linearly the Mueller parameters vector to the raw measured intensity vector. The multi-component nature of the Mueller matrices image is treated here by seeing the Mueller image as a result of a multi-channel imaging system where the information is conveyed to the user by a multi-channel image data. These channels will be indexed in the following, where appropriate, by jn∈ [0,15].

Regardless of the physical constraints that must be satisfied by the elements of , we have the same starting point as with the Stokes imaging framework [1]: the two problems share the same observation model. Systematic calibration errors as well as noises that arise from the use of a CCD photosensitive device as a detector (readout noise, thermal noise, dark current noise, and photon noise among others) are accounted for in the same way as in [1], which leads to the complete observation model:

I¯(jy,jx)=(1+ρ)P(η)M¯(jy,jx)+ε(jy,jn),

where ρ accounts for kinds of noises and errors having their variances directly proportional to the measured intensity (see [1] for more details) while ε encompasses all remaining noise terms.

The problem is now to estimate from the observations while ensuring the physical admissibility of the retrieved Mueller operators attached to each pixel. One can easily understand that the classical approach that relies on the PMM pseudo-inverse matrix is not the right way to achieve this task. Indeed, it necessitates a pixel-wise Mueller validity post checking which is time consuming without giving a proper estimate for pathological pixels.

We propose a new parameterization of the problem, yielding an observation model with simple constraints that allows tackling the estimation in a Bayesian framework. This comes at the cost of nonlinearity. The parameterization is achieved by observing that there is a one-to-one correspondence between a bona fide Mueller matrix M and a Hermitian positive semidefinite matrix H called “system coherency matrix”

H=k,l=03mkl[σkσl],

where {σk}3 k=0 comprising the Pauli matrices and the identity matrix, are the elements for a basis for ℂ2×2 and mkl are the elements of M [3]. The linear equation (4) leads to the following expression:

M¯=TH¯,

where T is a 16×16 invertible complex transformation matrix relating the stacked Mueller matrix to the stacked system’s coherency matrix.

It was stated that M is physically acceptable if and only if H is positive semi-definite [3]. Furthermore, if H is positive semidefinite, it can be written in terms of its Cholesky decomposition as [4]

H=ΛΛt,

where Λ is a lower triangular matrix composed of 16 real parameters (λi)

Λ=[λ0000λ4+iλ5λ100λ10+iλ11λ6+iλ7λ20λ14+iλ15λ12+iλ13λ8+iλ9λ3].

It is now possible to parameterize the elements of by {λi} in Eq. (5) and report the expression of in the observation model of Eq. (3) to obtain

I¯(jy,jx)=(1+ρ)P(η)TH¯({λi};jy,jx)+ε(jy,jn).

We must note that the observation model is now a nonlinear function of {λi} and that {λj, j∈[0,3]} must obey a positivity constraint to respect the positive definiteness of H. The estimation problem turns now to the estimation of the real parameters {λi} that are used to construct H, hence yielding a physically acceptable Mueller matrix. We note also that generalization to the case where the matrix P depends on the site (jy, jx) is straightforward.

2.2. Underlying segmentation model

The underlying segmentation map will be modeled as a Markov random field, and more precisely using Pott’s model [5]. Let

I0={I0(jy,jx);I0(jy,jx)[1,K],jy[1,n],jx[1,m]}

denote the segmentation image, where K is the number of classes. The prior probability associated to I 0 reads

p(I0|β)=1Z(β)exp(βU(I0)),

where

U(I0)=t~s𝕀(I0(t)=I0(s))

is the abstract energy measure associated to the field and Z the partition function. Variables t and s denote pixel sites of the image, the ~ denotes neighboring sites (we used eight neighbors in the present work) and 𝕀(·) is the indicator function taking values 1 or 0 whether the condition between parentheses is true or false. This model favors piecewise constant images, the size of the patches being governed by the abstract inverse temperature parameter β. A key problem in using Eq. (10) as a penalty term is the computation of Z as a function of β. We rely here on thermodynamic integration, i.e., on the relation

dlogZdβ=1ZdZdβ=I0U(I0)exp(βU(I0))Z=E[U(I0)],

where the expectation is taken with respect to p(I 0|β).

From a theoretical point of view, we have

logZ=nmlogK+0βE[U(I0)]dβ.

From a practical point of view, values of E [-U (I 0)] are computed using Markov chain Monte Carlo approximations for different values of β. The corresponding set of points is linearly interpolated, thus defining an approximation of dlog Z/dβ [6] which is further integrated thus giving an approximation of logZ. Let us mention that the samples of I 0 are generated by alternating Gibbs sampling (pixel-flipping) raster scans of the image with Swendsen-Wang iterations (cluster-flipping). The role of the latter iterations is to improve the mixing properties of the Markov chain. We will not further detail this standard procedure, the reader being referred to [5].

Let Φ be the parameter matrix defined as

Φ=[λ0(1)λ0(K)λ15(1)λ15(K)],

where the numbers between parentheses stand for the class index. In the estimation procedure to be described, the role of the segmentation map I 0 will be to index the appropriate column k of candidate Mueller class parameters {λi (k)} for pixels labeled as class k: I 0 (jy, jx)=k. This straightforwardly yields the vector attached to such pixel.

2.3. Noise model

The noise will be supposed to be independent and Gaussian distributed according to

ε(jy,jx,jn)I0(jy,jx)~𝒩(μI0(jy,jx)(jn),(σI0(jy,jx)(jn))2),

where (µ (jn) k,σ (jn) k) are the noise parameters (mean and standard deviation) that affect the kth class in the jn channel image.

Let us define Θ as the set Θ={Θ(jn); jn∈[0,15]}, where

Θ(jn)=[μ1(jn)μ2(jn)μK(jn)σ1(jn)σ2(jn)σK(jn)].

All these parameters are unknown and will have to be estimated. This general model will in the sequel be restricted to the case of vanishing means. From a modelling point of view, the model with non vanishing means suffers from an identifiability problem: modifying some or all the λi’s amounts to modifying the µj’s and we have, at least for now, no information (prior or measured) to be able to make a decision. Furthermore, this assumption does not penalize at all the algorithm performances as observed in practice.

2.4. General probabilistic model

The overall probabilistic model may be depicted using a Directed Acyclic Graph (DAG) (see Fig. 1).

For the sake of clarity, we drop the dependency on K, since the only role of K is in defining dimensions. Considering the DAG, we may write:

p(I,Φ,I0|ρ,Θ,β,σλ)=p(I|Φ,I0,ρ,Θ)p(Φ|σλ)p(I0|β).

The component terms on the right-hand-side are reckoned as follows:

p(I 0|β) is established by Eq. (10) and section 2.2;

 figure: Fig. 1.

Fig. 1. The directed acyclic graph representing the dependence relations between the variables. According to usual conventions, square boxes represent fixed or observed quantities whereas circles represent quantities to be estimated. More precisely, β is the abstract inverse temperature parameter of the Markov random field,σλ is the standard deviation of the λi’s, ρ is related to the quantum efficiency of the camera, I 0 is the underlying class segmentation map, Φ is the Mueller parameter matrix encompassing the λi(k)’s, Θ is the matrix encompassing the parameters of the noise distribution, and I is the intensity observations matrix.

Download Full Size | PDF

p(Φ|σλ) is defined in a class-independent manner as follows:

{λi~𝒩(0,σλ2)𝕀(λi0)i[0,3],λi~𝒩(0,σλ2)i[4,15].

The variance σ 2 λ is chosen sufficiently large in order to have no influence. In other words, the prior p(Φ|σλ) is flat compared to the likelihood p(I|Φ,I 0,ρ,Θ) when the λi(k) vary. Thus the prior p(Φ|σλ) has no influence besides imposing the positivity of λ0,⋯λ 3 when estimating Φ. The variance σ 2 λ may be estimated using physical considerations or using a set of I’s (from this set of I’s, we compute the H’s in a pseudo-inverse framework and estimate the λi’s by minimizing ‖H-ΛΛt2). This would give a set of λi’s from which we could derive σ 2 λ;

• the data-dependent term p(I|Φ,I 0,ρ,Θ) is obtained by solving Eq. (8) for ε then using Eq. (15) to express the (joint) probability density function of ε, assuming iid of pixels I(jy, jx) conditioned on I 0:

{p(IΦ,I0,ρ,Θ)=p(ε)=p(I¯(1+ρ)P(η)TH¯({λi}))=jy,jx,jn𝒩(ζ(jy,jx,jn),(σI0(jy,jx)(jn))2),

where

ζ(jy,jx,jn)=I(jy,jx,jn)(1+ρ)P(η)TH¯({Φ[I0(jy,jx)]},jy,jx)[jn].

In other words, the λi(k) are indexed from Φ (per Eq. (14)) based on the class index of the pixel, which in turn is looked-up from the segmentation map I 0.

3. Estimation procedure

All available information is contained in distribution (17), from which we will derive the estimates Φ^ , Î 0, ρ^,Θ^ ,β^. Let us emphasize that this method is fully unsupervised: there is no free parameter to be chosen or tuned by the user, set aside the number K of classes (see discussion below).

Let {ρ,Θ,β} be the set of hyperparameters (i.e., the parameters which are not directly of interest), whereas the parameters are I 0 and Φ. The parameters and hyperparameters will be obtained simultaneously by joint optimization of distribution (17), thus corresponding to generalized maximum likelihood estimation whereas maximum a posteriori would correspond to optimizing the distribution with respect to the parameters for given values of the hyperparameters.

To address the optimization problem, we resort to the Gauss-Seidel scheme: in the global loop, we consider all variables in a row and optimize -log p(I,Φ, I 0|ρ,Θ,β,σλ) with respect to the current variable while the other variables are held fixed. The update of I 0 is achieved site by site in a raster scan. The update of Φ is achieved component by component (i.e., we loop on i and k and optimize with respect to λi (k)). Convergence is reached when I 0 is not modified by an entire raster scan (in fact, we will monitor that I 0 is held constant over several (10 in the present case, see the counter<10 condition in Algorithm 2-b) raster scans, since even if I 0 is held constant, the other variables may vary, if only a little, which in turn may induce a modification of I 0).

Optimization with respect to λi (k) deserves particular comments. This optimization involves a fourth degree polynomial obtained by extracting the corresponding terms from -log p(I|Φ,I 0,ρ,Θ)-log p(Φ|σλ). This is handled using formal computation: the coefficients of the polynomial are analytically precomputed and the local minima are computed given the numerical values of the coefficients. The updated value of λi (k) is sought among a) the arguments of the local minima when no positivity constraint is involved or b) the positive arguments of the local minima and zero when a positivity constraint is involved.

The present optimization problem being highly multimodal, particularly with respect to I 0 when the noise level is important, we suggest using the algorithm detailed below which behaves better than straightforward and sole Gauss-Seidel descent.

At first, twenty different runs of the K-means clustering procedure are applied to the image set of m×n dimensional intensity channel vectors, yielding different class segmentation maps. The best resulting map, characterized by minimal distances between the clusters and the class’s centroids, is retained as the initial value of I 0 and used to initialize the DAG with the procedure detailed in Algorithm 1-a and Algorithm 1-b. In this phase, a whole scan of I 0 may lead to numerous modifications since the state of the DAG is far from its final (equilibrium) state. We want to avoid such a large movement without updating the rest (i.e., all variables but I 0) of the DAG, since it may lead to local optima far from the solution. Thus, we update the other variables of the DAG as soon as a given (reduced) number of site updates are achieved (Algorithm 1-a). This allows reaching a state much closer to the equilibrium than the initial state. We then switch to entire raster scans (Algorithm 1-b) on which we loop until convergence. The role of the mask is to improve the speed of the procedure. The new mask (see the instruction update the mask) is composed of all sites that have just been modified and of the neighbors of those sites. At this stage, the state is not optimal: a better (in the sense of a lower criterion value) solution may be found. This state cannot be reached by updating single sites at a time. This is why we resort to a multiresolution optimization scheme (Algorithm 2-a), where the sites of I 0 are modified blockwise. The other parameters of the DAG remain fixed in this phase, since they are close to their final values. We conclude with a standard Gauss-Seidel procedure (Algorithm 2-b) until convergence.

 figure: Algorithm 1:

Algorithm 1: Optimization phase 1: initialization.

Download Full Size | PDF

4. Algorithm application

4.1. Simulation results

The aim of this section is to compare the estimated Mueller images resulting from different approaches (K-means, pseudo-inverse, Bayesian) using a ground-truth. We synthesized a non trivial 128-pixels×128-pixels Mueller image by using a four class label map shown in Fig. 2. A different Mueller matrix was assigned to each label as shown in Table 1. For the Bayesian approach, we considered two Markov random field models: Potts’ model and the Markov mesh (unilateral model) considered in the previous work addressing the Stokes imaging case.

The PMM of an optimal rotating-retarder Mueller polarimeter using the optimal parameter vector ηopt as defined in Ref. [2] was used to generate the corresponding intensity images by mean of Eq. (2). The hypothesis and arguments about the image noise nature introduced in our previous paper [1] are still valid here. Thus we assume that ε follows a centered Gaussian distribution: a zero mean white Gaussian noise with variance σ 2 n was added to each intensity image for this simulation providing the necessary inputs to our model, i.e. intensity images as well as the PMM P. We present here simulations results carried out with two different noise variances (σ 2 n=0.1 and σ 2 n=0.5) which correspond respectively to signal-to-noise ratios over the mean transmittance image (first Mueller channel) about 10 dB and 3 dB.

 figure: Algorithm 2:

Algorithm 2: Optimization phase 2: refinement. In algorithm (a), all pixels of a given block are forced to the same label.

Download Full Size | PDF

The 16 channels of the Mueller image obtained by the classical pseudo-inverse solution are shown in Fig. 3. The results given by different approaches are summarized in Tables 2 and 3. Table 2 shows the percentage of misclassified pixels in the image. Table 3 presents the relative estimation error of the Mueller matrices associated to each class defined as

RE=MtrueMestimatedMtrue.

The Frobenius norm is used in the definition of RE.

We calculated also the mean RE value for the 3 dB case over the whole image for the classical

 figure: Fig. 2.

Fig. 2. 128-pixels×128-pixels mask used to generate the synthetic Mueller images. Each color (white, light gray, heavy gray and black) corresponds to one label.

Download Full Size | PDF

Tables Icon

Table 1. The four different matrices assigned to the classes of the label map of Fig. 2. Matrix (a) was assigned to the black pixels, matrix (b) to the heavy gray pixels, matrix (c) to the light gray pixels and matrix (d) to the white pixels in the label map

cal pseudo-inverse PDR approach as well as for the Bayesian one (Potts’ model) and obtained respectively 110% for the first one and 2.2% for the second one. When the Mueller matrix is estimated with the pseudo-inverse, 90%(resp. 99 %) of the pixels are not physically admissible for σ 2 n=0.1 (resp. σ 2 n=0.5). Moreover, for the pseudo-inverse estimate, several Mueller channels display no information (see for example column 3 and row 3 in Fig. 3), whereas structural information appears in all channels for the Bayesian estimate.

We conclude that the introduced Bayesian approach (Potts’ model) gives better results in terms of estimated quantities and structural information (composition) of the imaged scene as compared to other approaches (pseudo-inverse and K-means). Moreover, the Markov random field prior plays an important role (see Table 2, where the results significantly degrade for β=0). The two Bayesian variants (Potts’ model and the unilateral model) give nearly the same results in term of RE but Potts’ model is noticeably better in term of misclassified pixels (Table 2). Besides, further simulations showed that Potts’ model is more robust than the unilateral one with respect to increasing noise variance.

In real situations, for some limited scenario, the optimal number of classes (Kopt) can only be roughly estimated. Usually, it can’t be inferred easily by simple observation. Our approach is able to handle this issue. To determine Kopt, the optimum of the law in Eq. (17) is plotted against K (i.e., we plot -log(p(I,Φ^ ,Î 0|ρ^,Θ^ ,β^,σλ))+constant against K). We chose for Kopt the

 figure: Fig. 3.

Fig. 3. The 16 channels of the synthetic Mueller image obtained by the commonly used pseudo-inverse estimate obtained for the noise of variance σ 2 n=0.5.

Download Full Size | PDF

Tables Icon

Table 2. Mean percentage of misclassified pixels for the Bayesian approach (with (β^) and without (β=0) Markov random field prior) and for K-means clustering applied to the intensity images. Averaging has been done over noise realizations. Numbers between parentheses represents standard-deviation of related quantities.

value for which the graph levels off as shown in Fig. 4. In a fully Bayesian framework, we would have Kopt=argmaxK p(I|K). This would necessitate to resort to Markov chain Monte Carlo methods which is outside the scope of this paper. This is left for future work.

Running the method takes 4 minutes (resp. 30 minutes) for K=4 (resp. K=8) classes (128×128 image, 2.8 GHz standard desktop computer, programmation in C langage). This important variation in CPU time is due to an increase in the number of iterations.

4.2. Results against measured data

The proposed model is applied to real measured images acquired by an inline rotating quarter-wave plate Mueller imaging polarimeter. The object used to carry out these measurements consists of two dichroic patches (Polaroid) and a transparent thin film (cellophane) sandwiched between two coverglasses. The overall mount is imaged through a narrow band interferential

Tables Icon

Table 3. Mean values of the relative error RE (in percent) of the four Mueller matrices corresponding to the four classes of Fig. 2. Averaging has been done over noise realizations. The first values correspond to σ 2 n=0.1, the second values correspond to σ 2 n=0.5. The K-means were used to derive a segmentation map over which the pseudo-inverse estimate was class-wise averaged, yielding the k’s. We ensured physical admissibility conditions over those averages according to M¯̂k=TH¯̂k , with Ĥk=Λ̂kΛ̂kt and Λ̂k=argminΛM¯~kTΛΛ¯t2 , the minimization being achieved under the admissibility conditions.

 figure: Fig. 4.

Fig. 4. The optimal number of classes for the simulated data. The graph corresponds to σ 2 n=0.5.

Download Full Size | PDF

filter (10 nm) and a PSA in front of a 12 bits radiometric CCD camera. We point out that the dichroic patches are placed such that their transmission axes are nearly orthogonal. Totally polarized incident light undergoes multiple reflections between the two coverglasses and escapes highly depolarized such that the polarization signature is expected to be relatively weak. This situation was chosen to represent a hard test case that can be encountered in many practical situations such as in polarization imaging of biological tissues.

The Mueller image obtained by classical PDR of this object is shown in Fig. 5. At first, we can observe that while the conventional intensity imaging provides little to no information about the object (m 00 channel in the Mueller image), the other Mueller channels provide a rich enhancement to classical imaging. We observe also that the Mueller image is highly contaminated by a noise that hampers the physical interpretation of the interaction mechanisms which would allow inferring the object properties.

The algorithm output is given in Fig. 6 and in Table 4. We point out that all retrieved Mueller matrices are physically valid. This can be verified easily by computing the eigenvalues spectra of their corresponding coherency matrices. Indeed, all the eigenvalues are real positive. In the other hand, Fig. 7 shows the variation of the optimum of Eq. (17) versus the number of classes. The shape of this curve indicates that the optimal number of classes is here 5. This is in complete agreement with the a priori knowledge.

Once the Mueller matrices are estimated, one can focus on the physical interpretation of the

 figure: Fig. 5.

Fig. 5. The Mueller image of the test object as obtained from classical pseudo-inverse PDR. The image in the upper left (m 00) element corresponds to a conventional intensity image.

Download Full Size | PDF

 figure: Fig. 6.

Fig. 6. The final map obtained by the Bayesian approach (Potts’ model). Each gray level corresponds to one different class associated to a different Mueller matrix. The circular cross section delimits the footprint of the illuminating beam. Classes (2) and (3) are the dichroic patches and class (4) corresponds to the thin cellophane film.

Download Full Size | PDF

Tables Icon

Table 4. The five retrieved Mueller matrices that correspond to the label map output of our algorithm. These matrices have to be interpreted with the class numbering that is shown in Fig. 6.

 figure: Fig. 7.

Fig. 7. The optimal number of classes for the true image. We observe that the graph is consistent with the a priori knowledge of the object.

Download Full Size | PDF

results. Several analysis tools are available in the literature. We cite among others the coherency matrix decomposition approach [3] and the polar decomposition [7]. For this image we observe that the five obtained matrices depolarize a large amount of any incoming polarized state which is in agreement with the physical situation. We can also compute the product of the two matrices M 2 and M 3 corresponding to the second and third classes (which are weak polarizers oriented orthogonally to each other) to obtain

M2M3=[10.030.010.000.050.000.000.000.020.000.000.000.000.000.000.00],

which is very close to a pure depolarizer in accordance with the physics of the object (transmitting axes nearly orthogonal).

Finally, let us mention that running the algorithm takes 2 hours for K=4 classes (256×256 image, 2.8 GHz standard desktop computer, programmation in C langage).

5. Conclusion and perspectives

In this paper, we present a polarimetric data reduction method addressing piecewise constant images in the Mueller framework imagery. The proposed Bayesian method enforces the physical admissibility conditions while providing a robust clustering of the estimated images. The results obtained are significantly better and more robust than those given by classical pseudoinverse or K-means estimation. This framework will enable the researcher to enhance and deepen the understanding of the object constituents. Extension of the approach to smoothly varying polarization signatures is under investigation.

The software that processes Stokes and Mueller images is available on request for interested readers wishing to process their own images.

Acknowledgement

The authors are indebted to two anonymous reviewers for helpful comments and suggestions and acknowledge the financial support of the Agence Nationale de la Recherche (ANR Polarimetry).

References and links

1. J. Zallat and Ch. Heinrich, “Polarimetric data reduction: a Bayesian approach,” Opt. Express 15, 83–96 (2007), http://www.opticsinfobase.org/abstract.cfm?URI=oe-15-1-83. [CrossRef]   [PubMed]  

2. J. Zallat, S. Aïnouz, and M.-P. Stoll, “Optimal configurations for imaging polarimeters: impact of image noise and systematic errors,” J. Opt. A: Pure Appl. Opt. 8, 807–814 (2006). [CrossRef]  

3. S. Cloude and E. Pottier, “Concept of polarization entropy in optical scattering: polarization analysis and measurement,” Opt. Eng. 34, 1599–1610 (1995). [CrossRef]  

4. G. Golub and C. Van Loan, Matrix computation (The Johns Hopkins U. Press, Third edition, 1996).

5. G. Winkler, Image analysis, random fields and Markov chain Monte Carlo methods: a mathematical introduction (Springer, Second edition, 2006).

6. P. Green and S. Richardson, “Hidden Markov models and disease mapping,” J. Am. Stat. Assoc. 97, 1055–1070 (2002). [CrossRef]  

7. S.-Y. Lu and R. Chipman, “Interpretation of Mueller matrices based on polar decomposition,” J. Opt. Soc. Am. A 13, 1106–1113 (1996). [CrossRef]  

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (9)

Fig. 1.
Fig. 1. The directed acyclic graph representing the dependence relations between the variables. According to usual conventions, square boxes represent fixed or observed quantities whereas circles represent quantities to be estimated. More precisely, β is the abstract inverse temperature parameter of the Markov random field,σλ is the standard deviation of the λi ’s, ρ is related to the quantum efficiency of the camera, I 0 is the underlying class segmentation map, Φ is the Mueller parameter matrix encompassing the λi (k)’s, Θ is the matrix encompassing the parameters of the noise distribution, and I is the intensity observations matrix.
Algorithm 1:
Algorithm 1: Optimization phase 1: initialization.
Algorithm 2:
Algorithm 2: Optimization phase 2: refinement. In algorithm (a), all pixels of a given block are forced to the same label.
Fig. 2.
Fig. 2. 128-pixels×128-pixels mask used to generate the synthetic Mueller images. Each color (white, light gray, heavy gray and black) corresponds to one label.
Fig. 3.
Fig. 3. The 16 channels of the synthetic Mueller image obtained by the commonly used pseudo-inverse estimate obtained for the noise of variance σ 2 n =0.5.
Fig. 4.
Fig. 4. The optimal number of classes for the simulated data. The graph corresponds to σ 2 n =0.5.
Fig. 5.
Fig. 5. The Mueller image of the test object as obtained from classical pseudo-inverse PDR. The image in the upper left (m 00) element corresponds to a conventional intensity image.
Fig. 6.
Fig. 6. The final map obtained by the Bayesian approach (Potts’ model). Each gray level corresponds to one different class associated to a different Mueller matrix. The circular cross section delimits the footprint of the illuminating beam. Classes (2) and (3) are the dichroic patches and class (4) corresponds to the thin cellophane film.
Fig. 7.
Fig. 7. The optimal number of classes for the true image. We observe that the graph is consistent with the a priori knowledge of the object.

Tables (4)

Tables Icon

Table 1. The four different matrices assigned to the classes of the label map of Fig. 2. Matrix (a) was assigned to the black pixels, matrix (b) to the heavy gray pixels, matrix (c) to the light gray pixels and matrix (d) to the white pixels in the label map

Tables Icon

Table 2. Mean percentage of misclassified pixels for the Bayesian approach (with ( β ^ ) and without (β=0) Markov random field prior) and for K-means clustering applied to the intensity images. Averaging has been done over noise realizations. Numbers between parentheses represents standard-deviation of related quantities.

Tables Icon

Table 3. Mean values of the relative error RE (in percent) of the four Mueller matrices corresponding to the four classes of Fig. 2. Averaging has been done over noise realizations. The first values correspond to σ 2 n =0.1, the second values correspond to σ 2 n =0.5. The K-means were used to derive a segmentation map over which the pseudo-inverse estimate was class-wise averaged, yielding the k ’s. We ensured physical admissibility conditions over those averages according to M ¯ ̂ k = T H ¯ ̂ k , with H ̂ k = Λ ̂ k Λ ̂ k t and Λ ̂ k = arg min Λ M ¯ ~ k T ΛΛ ¯ t 2 , the minimization being achieved under the admissibility conditions.

Tables Icon

Table 4. The five retrieved Mueller matrices that correspond to the label map output of our algorithm. These matrices have to be interpreted with the class numbering that is shown in Fig. 6.

Equations (22)

Equations on this page are rendered with MathJax. Learn more.

I ( j y , j x ) = A M ( j y , j x ) W ,
I ¯ ( j y , j x ) = ( W t A ) M ¯ ( j y , j x ) = P ( η ) M ¯ ( j y , j x ) ,
I ¯ ( j y , j x ) = ( 1 + ρ ) P ( η ) M ¯ ( j y , j x ) + ε ( j y , j n ) ,
H = k , l = 0 3 m k l [ σ k σ l ] ,
M ¯ = T H ¯ ,
H = Λ Λ t ,
Λ = [ λ 0 0 0 0 λ 4 + i λ 5 λ 1 0 0 λ 10 + i λ 11 λ 6 + i λ 7 λ 2 0 λ 14 + i λ 15 λ 12 + i λ 13 λ 8 + i λ 9 λ 3 ] .
I ¯ ( j y , j x ) = ( 1 + ρ ) P ( η ) T H ¯ ( { λ i } ; j y , j x ) + ε ( j y , j n ) .
I 0 = { I 0 ( j y , j x ) ; I 0 ( j y , j x ) [ 1 , K ] , j y [ 1 , n ] , j x [ 1 , m ] }
p ( I 0 | β ) = 1 Z ( β ) exp ( β U ( I 0 ) ) ,
U ( I 0 ) = t ~ s 𝕀 ( I 0 ( t ) = I 0 ( s ) )
d log Z d β = 1 Z d Z d β = I 0 U ( I 0 ) exp ( β U ( I 0 ) ) Z = E [ U ( I 0 ) ] ,
log Z = n m log K + 0 β E [ U ( I 0 ) ] d β .
Φ = [ λ 0 ( 1 ) λ 0 ( K ) λ 15 ( 1 ) λ 15 ( K ) ] ,
ε ( j y , j x , j n ) I 0 ( j y , j x ) ~ 𝒩 ( μ I 0 ( j y , j x ) ( j n ) , ( σ I 0 ( j y , j x ) ( j n ) ) 2 ) ,
Θ ( j n ) = [ μ 1 ( j n ) μ 2 ( j n ) μ K ( j n ) σ 1 ( j n ) σ 2 ( j n ) σ K ( j n ) ] .
p ( I , Φ , I 0 | ρ , Θ , β , σ λ ) = p ( I | Φ , I 0 , ρ , Θ ) p ( Φ | σ λ ) p ( I 0 | β ) .
{ λ i ~ 𝒩 ( 0 , σ λ 2 ) 𝕀 ( λ i 0 ) i [ 0 , 3 ] , λ i ~ 𝒩 ( 0 , σ λ 2 ) i [ 4 , 15 ] .
{ p ( I Φ , I 0 , ρ , Θ ) = p ( ε ) = p ( I ¯ ( 1 + ρ ) P ( η ) T H ¯ ( { λ i } ) ) = j y , j x , j n 𝒩 ( ζ ( j y , j x , j n ) , ( σ I 0 ( j y , j x ) ( j n ) ) 2 ) ,
ζ ( j y , j x , j n ) = I ( j y , j x , j n ) ( 1 + ρ ) P ( η ) T H ¯ ( { Φ [ I 0 ( j y , j x ) ] } , j y , j x ) [ j n ] .
RE = M true M estimated M true .
M 2 M 3 = [ 1 0.03 0.01 0.00 0.05 0.00 0.00 0.00 0.02 0.00 0.00 0.00 0.00 0.00 0.00 0.00 ] ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.