Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Y-Net: a dual-branch deep learning network for nonlinear absorption tomography with wavelength modulation spectroscopy

Open Access Open Access

Abstract

This paper demonstrates a new method for solving nonlinear tomographic problems, combining calibration-free wavelength modulation spectroscopy (CF-WMS) with a dual-branch deep learning network (Y-Net). The principle of CF-WMS, as well as the architecture, training and performance of Y-Net have been investigated. 20000 samples are randomly generated, with each temperature or H2O concentration phantom featuring three randomly positioned Gaussian distributions. Non-uniformity coefficient (NUC) method provides quantitative characterizations of the non-uniformity (i.e., the complexity) of the reconstructed fields. Four projections, each with 24 parallel beams are assumed. The average reconstruction errors of temperature and H2O concentration for the testing dataset with 2000 samples are 1.55% and 2.47%, with standard deviations of 0.46% and 0.75%, respectively. The reconstruction errors for both temperature and species concentration distributions increase almost linearly with increasing NUC from 0.02 to 0.20. The proposed Y-Net shows great advantages over the state-of-the-art simulated annealing algorithm, such as better noise immunity and higher computational efficiency. This is the first time, to the best of our knowledge, that a dual-branch deep learning network (Y-Net) has been applied to WMS-based nonlinear tomography and it opens up opportunities for real-time, in situ monitoring of practical combustion environments.

© 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

1. Introduction

Tunable diode laser absorption spectroscopy (TDLAS) has served as a powerful species diagnostic tool in energy and power systems over the past decades for non-intrusive, highly sensitive and quantitative measurements of various gas properties [13]. Direct absorption spectroscopy (DAS) and wavelength modulation spectroscopy (WMS) are the two most common schemes for TDLAS sensing. In particular, WMS is more advantageous for situations with low absorbance, high pressure, or blended absorption features [4,5].

However, as a line-of-sight (LOS) technique, TDLAS has been limited to path-integrated sensing along the laser beam. Various approaches have been adopted to retrieve spatially-resolved distributions, including single-beam absorption thermometry [6,7], tomographic absorption spectroscopy (TAS) with multiple laser beams [8,9], and laser absorption imaging (LAI) with a camera [1012]. Among these approaches, tomographic absorption spectroscopy (TAS) has evolved into a well-established imaging technique from laboratory specialties to real-field combustion environments [8].

Tomography is a canonical inverse problem, such that a forward model (a set of projection measurements) needs to be inverted to estimate the spatial distribution of the quantity of interest (e.g., gas-phase tomography) [13]. TAS can be divided into linear [8,1416] and nonlinear [8,1721] modalities according to the basic mathematics defining the problem. Linear TAS, an implementation of classical tomography, records typically one or two absorption transitions with narrow-bandwidth diode lasers. Usually a large number of projections (i.e., a large number of laser beams) are required to alleviate the ill-posedness of the linear equation system and achieve reasonable reconstruction. This largely restricts the application of linear TAS in practical combustion environments where optical access is highly limited. For nonlinear tomographic absorption spectroscopy (NTAS), hyperspectral laser sources [17] or multiplexed techniques [18] are typically adopted, constituting a set of nonlinear equations from which gas property distribution can be recovered. The additional spectral or modulation harmonics information in NTAS complements the reduction of laser beam paths and therefore effectively mitigate the ill-posedness of the inverse problem in linear TAS. The current work focuses on the demonstrations with NTAS.

Extensive research has been dedicated to NTAS using DAS or calibration-free WMS (CF-WMS) technique for the reconstruction of temperature and species concentration [8,1721]. The DA scheme is commonly used in NTAS due to relatively simple tomographic inversion from linearly aggregated absorbance along the LOS. Ma et al. exploited hyperspectral laser source for tomographic imaging using DAS technique, with a significant reduction in the number of required projections and enhanced stability of the reconstruction [17]. Recently, Cai et al. proposed a multiplexed absorption tomography technique with CF-WMS for simultaneous imaging of temperature and species concentration in harsh environments [18]. Different from linear TAS, NTAS using DAS and CF-WMS typically leads to a sizable system of nonlinear equations. Conventional tomographic algorithms developed for the linear TAS problems cannot solve the inverse problem of nonlinear tomography. It has been proven that the nonlinear tomographic problems can be converted to an optimization problem, and then solved by heuristic algorithms such as the simulated annealing (SA) algorithm [8], or evolution strategy based on covariance matrix adaption algorithm (CMA-ES) [21]. However, the performance of the aforementioned algorithms relies on problem-specific knowledge in heuristics and can also be difficult to implement due to the high computational cost.

Over the past few years, deep learning has gained a tremendous amount of attention for image restoration and reconstruction tasks [22,23], which coincidentally offers a promising alternative for inverse ill-posed problems. A dual-branch convolutional network, namely, CSTNet, was developed by Jiang et al. for high-fidelity, rapid, and simultaneous imaging of temperature and species concentration in linear tomographic problems [16]. For solving nonlinear tomographic problems, Huang et al. proposed a new inversion method based on convolutional neural network (CNN) and the results show that temperature distribution can be rapidly reconstructed using CNN [19]. Deng et al. further investigated the performance of other sophisticated deep neural networks such as deep belief networks (DBN) and recurrent neural network (RNN) in solving nonlinear NTAS problems [20]. However, only preliminary study of DAS-based deep learning methods for nonlinear tomographic problems has been demonstrated, and the reported deep learning networks can only be used to reconstruct either temperature or species concentration alone. Under the inspiration of the pioneering study of DAS-based deep learning method in linear TAS problems by Jiang et al. [16], it is appealing to investigate the nonlinear tomographic problems with more advanced deep learning network. In view of the diagnostics needs and progress, we here propose a novel nonlinear absorption tomography technique, combining a dual-branch deep learning network (Y-Net) with CF-WMS to simultaneously reconstruct temperature and species concentration in harsh environments. This is the first time, to the best of the authors’ knowledge, that a deep learning network is applied to WMS-based nonlinear tomography for simultaneous reconstruction of temperature and species concentration.

2. WMS-based nonlinear tomography

The principle of wavelength modulation spectroscopy has been well documented [15], and only relevant spectroscopic fundamentals are briefly reviewed here. In scanned-WMS scheme, a high-frequency sinusoidal modulation at frequency fm is used to modulate the laser wavelength around the targeted absorption transition. The instantaneous optical frequency and intensity can be expressed as:

$$v(t )= \bar{v} + {a_m}\cos ({2\pi {f_m}t} )$$
$${I_0}(t )= \overline {{I_0}} [{1 + {i_0}\cos ({2\pi {f_m}t + {\varphi_1}} )+ {i_2}\cos ({4\pi {f_m}t + {\varphi_2}} )} ]$$
where $\bar{v}{\kern 1pt} (\textrm{c}{\textrm{m}^{ - 1}})$ is the center optical frequency and am (cm-1) is the optical frequency modulation depths. i0 and i2 refer to the linear and nonlinear intensity modulation amplitudes (both normalized by the average laser intensity $\overline {{I_0}}$). ${\varphi _1}$ and ${\varphi _2}$ denote the linear and nonlinear phase shifts between the laser intensity modulation and frequency modulation.

The transmitted laser intensity It(t) can be obtained by the Beer-Lambert relation.

$${I_t}(t )\textrm{ = }G{I_0}(t )\cdot \tau [{v(t )} ]= G{I_0}(t )\cdot \sum\limits_{k = 0}^\infty {{H_k}({v,{a_m}} )} \cos ({2\pi {f_m}t} )$$
where G is the opto-electronic gain and Hk denotes the kth-order harmonic coefficient of the spectral transmissivity $\tau [v(t)].$ Harmonic WMS signals can be extracted from the transmitted intensity It(t) through digital lock-in detection. The magnitude of the first and second harmonic of the transmitted intensity It(t), termed S1f and S2f, can be described as:
$$\begin{array}{l} {S_{1f}}\textrm{ = }\frac{{G\overline {{I_0}} }}{2}\left\{ {{{\left[ {{H_1} + {i_0}\left( {{H_0} + \frac{{{H_2}}}{2}} \right)\cos ({{\varphi_1}} )+ \frac{{{i_2}}}{2}({{H_1} + {H_3}} )\cos ({{\varphi_2}} )} \right]}^2}} \right.\\ {\left. {{\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} + {{\left[ {{i_0}\left( {{H_0} - \frac{{{H_2}}}{2}} \right)\textrm{sin}({{\varphi_1}} )+ \frac{{{i_2}}}{2}({{H_1} - {H_3}} )\textrm{sin}({{\varphi_2}} )} \right]}^2}} \right\}^{1/2}} \end{array}$$
$$\begin{array}{l} {S_{2f}}\textrm{ = }\frac{{G\overline {{I_0}} }}{2}\left\{ {{{\left[ {{H_2} + \frac{{{i_0}}}{2}({{H_1} + {H_3}} )\cos ({{\varphi_1}} )+ {i_2}\left( {{H_0} + \frac{{{H_4}}}{2}} \right)\cos ({{\varphi_2}} )} \right]}^2}} \right.\\ {\left. {{\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} + {{\left[ {\frac{{{i_0}}}{2}({{H_1} - {H_3}} )\textrm{sin}({{\varphi_1}} )+ {i_2}\left( {{H_0} - \frac{{{H_4}}}{2}} \right)\textrm{sin}({{\varphi_2}} )} \right]}^2}} \right\}^{1/2}} \end{array}$$

By normalizing S2f by S1f, G and $\overline {{I_0}}$ can be eliminated, enabling calibration-free WMS-2f/1f scheme. As illustrated in Fig. 1, the tomographic field (i.e., the region of interest, ROI) is discretized into $D \times D$ square pixels. A number of laser diodes with center output frequency ${v_l}(l = 1,2, \ldots L)$ are combined using a wavelength multiplexer and each is modulated at a different frequency fm,l. The lasers are then split into multiple beams to traverse the ROI and then received by the detectors. As illustrated in [8], three types of beam arrangements including parallel-beam, fan-beam and irregular-beam arrangements, are typically implemented for tomographic schemes. Fan-beam arrangement usually suffers lower energy density than line illumination in a parallel-beam arrangement, while irregular-beam arrangement is typically preferred in practical scenarios where optical access is extremely limited. Here we choose to implement parallel-arrangement, and without loss of generality, each set of parallel beams is specified by an angle $\alpha ,$ a distance d, and multiplexed frequencies $({v_1},{v_2}, \ldots {v_L}).$

 figure: Fig. 1.

Fig. 1. Schematic diagram of multiplexed nonlinear absorption tomography based on calibration-free wavelength modulation spectroscopy.

Download Full Size | PDF

The detected signals are then post-processed using lock-in detection to recover L nonlinear equations for S2f/1f along each measurement projection. The sinogram ${[{({S_{2f/1f}})_v}]_{L \times M}}$ can be expressed as in Eq. (6), with M the number of parallel beams.

$${[{{{({{S_{2f/1f}}} )}_v}} ]_{L \times M}}\textrm{ = }{\left[ {\begin{array}{cccc} {{{({{S_{2f/1f}}} )}_{{v_1},1}}}&{{{({{S_{2f/1f}}} )}_{{v_1},2}}}& \cdots &{{{({{S_{2f/1f}}} )}_{{v_1},M}}}\\ {{{({{S_{2f/1f}}} )}_{{v_2},1}}}&{{{({{S_{2f/1f}}} )}_{{v_2},2}}}& \cdots &{{{({{S_{2f/1f}}} )}_{{v_1},M}}}\\ \vdots & \vdots & \ddots & \vdots \\ {{{({{S_{2f/1f}}} )}_{{v_L},1}}}&{{{({{S_{2f/1f}}} )}_{{v_L},2}}}& \cdots &{{{({{S_{2f/1f}}} )}_{{v_L},M}}} \end{array}} \right]_{L \times M}}$$

In this work, six H2O absorption lines located between 7294.0 cm-1 and 7444.4 cm-1 are chosen for H2O sensing according to Ref. [24]. The spectroscopic parameters for the targeted H2O lines provided by HITRAN 2020 database [25] including wavenumber v0 (cm-1), linestrength S (cm-2·atm-1), and lower-state energy E'‘ (cm-1) are listed in Table 1.

Tables Icon

Table 1. Spectroscopic data for the H2O transitions of interest taken from HITRAN 2020

Figure 2 shows the calculated S2f/1f at 7444.37 cm-1 under different temperature and H2O concentration conditions ${a_m},{i_0},{\varphi _1},{i_2},{\varphi _2}$ are set to $0.06{\kern 1pt} {\kern 1pt} \textrm{c}{\textrm{m}^{ - 1}},0.50,1.2\pi ,0.015$ and $1.4\pi$ respectively, according to typical near-infrared diode laser characteristics [4]. The modulation depth am may be further optimized to maximize 2f/1f signal intensity, but is not the main focus of this work. Unlike the DAS absorbance that varies linearly with concentration under optically thin conditions, the calculated WMS-2f/1f signal changes nonlinearly with both temperature and H2O concentration. This will further increase the complexity in the nonlinear tomographic inversion when using conventional heuristic algorithms.

 figure: Fig. 2.

Fig. 2. Calculated WMS-2f/1f signal with different temperature and H2O concentration, respectively.

Download Full Size | PDF

With 24 parallel beams used, the ROI is meshed into $24 \times 24$ pixels at four projection angles $(\alpha \textrm{ = }{\textrm{0}^\circ },{45^\circ },9{\textrm{0}^\circ },{135^\circ })$ resulting in a total number of 96 LOS measurement projections. The “true” field to be reconstructed is represented by assumed phantoms, each featuring three randomly positioned Gaussian distributions to characterize the multi-modal nature of the temperature or H2O concentration distributions in practical combustion environments [16]. The distributions of temperature and H2O concentration can be mathematically expressed as:

$$T({x,y} )\textrm{ = }{T_{\min }} + \frac{1}{3}\sum\limits_{i = 1}^3 {{\xi _i}({{T_{\max }} - {T_{\min }}} )} \exp \left[ { - \frac{{{{({x - {x_{c,i}}} )}^2} + {{({y - {y_{c,i}}} )}^2}}}{{2\sigma_T^2}}} \right]$$
$$X({x,y} )\textrm{ = }{X_{\min }} + \frac{1}{3}\sum\limits_{i = 1}^3 {{\xi _i}({{X_{\max }} - {X_{\min }}} )} \exp \left[ { - \frac{{{{({x - {x_{c,i}}} )}^2} + {{({y - {y_{c,i}}} )}^2}}}{{2\sigma_X^2}}} \right]$$
where $({x_{c,i}},{y_{c,i}})$ with i = 1,2,3 denotes the center position of the ith Gaussian distribution along direction x and y. Here, ${x_{c,i}}$ or ${y_{c,i}}$ is randomly selected from interval [2,23]. Tmax (Xmax) and Tmin (Xmin) are the maximum and minimum temperature (H2O concentration), respectively. According to typical combustion flame conditions, Tmin, Tmax, Xmin and Xmax are set to 500 K, 2500 K, 0.02 and 0.12, respectively. ${\xi _i} \sim U(0.8,1.0)$ is a random scaling factor. ${\sigma _T}$ (cm) and ${\sigma _X}$ (cm), the standard deviations of temperature and H2O concentration distribution, represents the spatial spread of the Gaussian flame. The spread of temperature is caused by heat transfer and dissipation, while the spread of H2O concentration is a result of flow convection and molecular motion. Based on the physical understanding that the temperature and H2O concentration distributions of the axisymmetric flame are correlated, with the former being wider [16], it is reasonable to further assume ${\sigma _X} = \rho {\sigma _T}$ with $\rho \sim U(0.75,1.0)$ and ${\sigma _T}$ selected from interval [3,6].

The performance of tomographic reconstruction is closely related to the complexity of temperature and concentration distributions. However, quantitative characterization of such spatial non-uniformity is challenging and case-dependent. Recently, Xiao et al. proposed a novel approach, namely the non-uniformity coefficient (NUC) method based on uniform design theory (i.e., star discrepancy measure), to quantify the non-uniformity of temperature or concentration distributions [26]. It has been demonstrated that the NUC method outperforms existing methods such as the standard deviation (SD) method [27], the coefficient of variation (CoV) method [28], and the image entropy (IE) method [29], by taking into account the time space and position information (TSPI) of targeted distributions. Here in our work, NUC method offers a benchmark to quantify the non-uniformity (i.e., complexity) of the reconstructed temperature and species concentration fields in tomographic problems. Take temperature distribution as example, the temperature local discrepancy function (TLDF), ${\boldsymbol d}_{PQ}^ \ast ({\boldsymbol \theta }),$ of the local rectangular region ${\boldsymbol \theta }\textrm{ = [}0,{\theta _1}\textrm{]} \times [0,{\theta _2}]$ in the ROI shown in Fig. 1, can be expressed as:

$${\boldsymbol d}_{PQ}^ \ast ({\boldsymbol \theta } )\textrm{ = }\frac{{\sum\nolimits_{j = 1}^{{\theta _2}} {\sum\nolimits_{i = 1}^{{\theta _1}} {{T_{i,j}}} } }}{{\sum\nolimits_{j = 1}^D {\sum\nolimits_{i = 1}^D {{T_{i,j}}} } }}\textrm{ - }\frac{{{\theta _1}{\theta _2}}}{{{D^2}}}$$
where Ti,j denotes the temperature value at the discretized square pixel (i, j), ${\theta _1} \in [1,D], {\theta _2} \in [1,D]$ and $\frac{{{\theta _1}}}{{{\theta _2}}} = 1$ for the square region shown in Fig. 1. The non-uniformity coefficient (NUCq, q = 1,2,3,4) of the temperature field is given by:
$$NUC_q^T = \sup |{{\boldsymbol d}_{PQ}^ \ast ({{\boldsymbol \theta },q} )} |$$
where $|{{\boldsymbol d}_{PQ}^ \ast ({\boldsymbol \theta },q)} |$ denotes the absolute value of ${\boldsymbol d}_{PQ}^ \ast ({\boldsymbol \theta },q).$ In particular, $NUC_1^T,NUC_2^T, NUC_3^T$ and $NUC_4^T$ denote the non-uniformity characterizations from the top-left, bottom-left, bottom-right and top-right orientations in the whole ROI. Then, the overall non-uniformity of temperature field, NUCT, is defined as follows:
$$NUC_q^T = \max \{{NUC_1^T,NUC_2^T,NUC_3^T,NUC_4^T} \}$$

A large NUCT indicates temperature deviations from uniformity such as local heterogeneity or lop-sidedness, while small discrepancy indicates that the temperature is more uniform in the whole ROI. Non-uniformity coefficient of species concentration distribution is similarly denoted by NUCX.

As depicted in Fig. 3, two examples of representative phantoms with different multi-modal distributions of temperature and H2O concentration are demonstrated. The three Gaussian distributions in phantoms (a) and (b) overlap with each other, and the 96 LOS measurement projections of WMS-2f/1f at the six absorption transition peaks are shown in Fig. 3(c). NUC values of phantoms (a) and (b) are 0.096 and 0.105, and the corresponding standard deviations are 442.36 K and 0.022, respectively. Another example with three distributions separated further apart with different locations and sizes is depicted in phantoms (d) and (e) and the accumulated WMS-2f/1f is shown in Fig. 3 (f). Phantoms (d) and (e) present NUC values of 0.055 and 0.062, with standard deviations of 204.97 K and 0.010, respectively. The horizontal axes of Fig. 3 (c) and (f) correspond to the 24 LOS measurement projections at each projection angle.

 figure: Fig. 3.

Fig. 3. Two examples of representative phantoms with different multi-modal distributions of temperature and H2O concentration, and the corresponding WMS-2f/1f. Example 1: (a) ∼ (c), Example 2: (d) ∼ (f). The red crosses are at the center of the phantoms. Horizontal axes of Fig. (c) and (f) correspond to 24 LOS measurement projections at each projection angles, i.e., $(\alpha \textrm{ = }{\textrm{0}^\circ },{45^\circ },9{\textrm{0}^\circ },{135^\circ }),$ respectively.

Download Full Size | PDF

The resultant sinograms, i.e., the 96 LOS measurement projections in Fig. 3 (c) and (f), are closely related to the characteristics of temperature and H2O concentration distributions. As can be seen in Fig. 3 (c), the accumulated WMS-2f/1f appears more symmetric about the center of the ROI at all four projection angles when the three phantoms closely overlap with each other, whereas the example of Fig. 3 (f) exhibits more asymmetric and complexed features with the phantoms further separated apart. A key benefit of the deep learning methods is the automated extraction of complex data representations, making it a valuable tool especially when dealing with more complicated raw sinogram datasets. High-fidelity tomographic reconstructions can therefore be achieved with deep learning methods by recasting the reconstruction task into a data-driven problem and learning the intricate features from the large amount of tomography data.

3. Y-Net: a dual-branch deep learning network

Deep learning network offers a promising alternative for solving nonlinear tomographic inversion problems. However, as discussed in Section 1, early work by Jiang et al. have only explored such potential for linear tomographic problems [16]. For nonlinear tomographic problems, previous attempts to apply deep learning method to tomographic combustion diagnostics have not considered the inter-dependence of temperature and species concentration as reflected in the molecular absorption spectroscopy data, and only temperature or species concentration reconstruction alone with DAS-based nonlinear tomographic problems has been demonstrated. For the first time, we propose a dual-branch deep learning network (Y-Net) for nonlinear absorption tomography with CF-WMS considering the cross-correlation between temperature and species concentration distributions. Benefiting from the architecture of two up-sampling paths in Y-Net, the associated temperature and species concentration can be simultaneously reconstructed with high accuracy and low computational cost.

Y-Net is conceptually simple and generalizes the U-Net, and the latter has been proven to be an effective segmentation and classification network in biomedical images [3032]. Simplified U-Net architecture for 2-D signal implementation is illustrated in Fig. 4. The U-Net architecture consists of a contracting path and an expansive path, serving as the encoder and decoder paths, respectively. Here, the input ${I_{\textrm{in}}} \in {{\mathbb R}^{n \times n}}$ is first filtered by several repeated convolutional filters $\bar{\phi },$ followed by max pooling operation $\varphi \in {{\mathbb R}^{\frac{n}{2} \times n}}$ to generate down-sampled feature maps. Mathematically, the down-sampling process can be described as:

$${I_d} = \varphi ({{I_{\textrm{in}}} \otimes \bar{\phi }} )$$
where ${I_{\textrm{in}}} \otimes \bar{\phi }$ denotes the multi-channel convolution in U-Net. The max pooling operation $\varphi$ is given by:
$$\varphi \textrm{ = }\left[ {\begin{array}{ccccc} {{b_1}}&{1 - {b_1}}& \cdots &0&0\\ 0&0& \cdots &0&0\\ \cdots & \cdots & \ddots & \cdots & \cdots \\ 0&0& \cdots &{{b_{\frac{n}{2}}}}&{1 - {b_{\frac{n}{2}}}} \end{array}} \right] \in {{\mathbb R}^{\frac{n}{2} \times n}}$$
where ${\{{{b_i}} \}_{i = 1,2, \cdots ,\frac{n}{2}}}$ are random binary numbers (0, 1) determined by the signal statistics. The number of filter channels is doubled and the dimension of the input image is reduced by half after each max pooling operation. Each step in the expansive path is followed by an up-sampling step of the feature map to recover the original dimensions of the input images. Note that high-spatial-frequency information such as sharp object boundaries may be lost when employing consecutive pooling layers. Therefore, a skip connection with the correspondingly cropped feature map from the contracting path is used and the final recovered output can be described as:
$${I_{\textrm{out}}} = {I_{\textrm{in}}} \otimes \bar{\phi } + {\varphi ^T}\varphi ({{I_{\textrm{in}}} \otimes \bar{\phi }} )$$

 figure: Fig. 4.

Fig. 4. Simplified U-Net architecture.

Download Full Size | PDF

The Y-Net, an improved version of U-Net, adds a second parallel branch and exploits a dual-branch feature extraction module. Previous works based on a Y-shape-structured neural network have been demonstrated for poly detection [33], photoacoustic imaging [34], digital holographic reconstruction [35,36], etc. The proposed dual-branch Y-Net in this work, enables simultaneous reconstructions of both temperature and species concentration.

The global architecture of Y-Net is Sillustrated in Fig. 5, consisting of an encoder-decoder network with sum-skip-concatenation connections. The sinograms ${[{({S_{2f/1f}})_v}]_{6 \times 96}}$ obtained in Fig. 1 is reshaped into a matrix with size of $24 \times 24,$ serving as the input image of the proposed Y-Net. It is worth noting that variably shaped and sized images (sinograms) can be given as the input to the deep learning network with an updated Y-Net architecture.

 figure: Fig. 5.

Fig. 5. Global architecture of Y-Net in this work.

Download Full Size | PDF

The encoding network in Y-Net, i.e., the down-sampling path on the left side, contains three repeated stages to down-sample the feature maps. The number noted on the top (or bottom) of a block denotes the number of filter channels, i.e., the feature depth, and the number on the left indicates the size of the feature map. The arrows as explained in the legend indicate the operations between two layers. Each stage of the down-sampling path has double $3 \times 3$ convolution kernel with a stride of 1, followed by a batch normalization (BN) and a rectified linear unit (ReLU). A $2 \times 2$ max pooling operation layer with a stride of 2 is applied between the stages in the down-sampling path. As illustrated in Fig. 5, the number of filter channels is doubled after each stage of down-sampling, with the dimension of the input image reduced by half each time. The bridge path is similar to the down-sampling path. The decoding network in Y-Net on the right side consists of two up-sampling paths for either temperature or H2O concentration inference. The skip connection concatenates the feature maps from down-sampling path to up-sampling path, providing an alternative way to ensure feature reusability and recover spatial information loss during down-sampling. Each stage in the up-sampling paths includes repeated and transposed convolution layers to up-sample the feature maps, followed by concatenations from the down-sampling path and double $3 \times 3$ convolutions. Finally, with a fully connected layer, the $576 \times 1$ feature vector is converted into $24 \times 24$ temperature or H2O concentration distribution. The proposed Y-Net has a total of 87 layers and 24,052,512 (24 M) learnable parameters including weights and bias at each layer.

The total mean square error (MSE) loss function to evaluate the reconstruction error of both temperature and H2O concentration is defined as:

$${L_{TX}} = \lambda {L_T} + {L_X}$$
where LT and LX denote the MSE of temperature and species concentration reconstruction, respectively, and $\lambda$ is a weighting hyper-parameter. LT and LX can be expressed as:
$${L_T} = \frac{1}{N}\sum\limits_n^N {{{||{T_n^G - T_n^R} ||}^2}}$$
$${L_X} = \frac{1}{N}\sum\limits_n^N {{{||{X_n^G - X_n^R} ||}^2}}$$
where $T_n^G(X_n^G)$ and $T_n^R(X_n^R)$ are the normalized ground truth (G) and reconstruction (R) of temperature (H2O concentration) distributions, respectively. n is the number of the training dataset, and N is the mini-batch size.

4. Results and discussions

4.1 Training of Y-Net

A total number of 20000 samples are randomly generated, of which 80% are used for training, 10% for validation, and 10% for testing. Each sample consists of a temperature phantom, a species concentration phantom and a sinogram with the dimension of $6 \times 96.$ As discussed in Section 2, the complexity of the reconstructed temperature and species concentration distributions in this work is quantitively characterized by the NUC method. The histograms in Fig. 6 (a) and (b) show the NUC value distributions of 20000 temperature and species concentration phantoms, respectively. The average NUC value of 20000 temperature phantoms is 0.0887 and the standard deviation is 0.0266. 20000 species concentration phantoms present an average NUC value of 0.0971 with a standard deviation of 0.0294.

 figure: Fig. 6.

Fig. 6. NUC value distributions of 20000 temperature (a) and species concentration (b) phantoms, respectively.

Download Full Size | PDF

MATLAB R2021a environment (Mathwork, Natick) is used to implement the proposed Y-Net. The training of the network is performed on a PC with Core i9-10900 K CPU (3.7 GHz) and 32 GB of RAM, using NVIDIA GeForce GTX 1050 Ti GPU. Adaptive moment estimation (ADAM) algorithm is chosen to perform the optimization [37]. The initial learning rate is set to 0.0002 determined by range test [19,20] and halved every 10 epochs. The network is trained with a total epoch of 20 and a mini-batch size of 32. To prevent over-fitting of the neural network, the training stops when the network performance on the validation dataset begins to deteriorate. $\lambda$ is set to 1, indicating that the reconstruction of H2O concentration is of equal importance to that of temperature. Figure 7 shows the evolution of the loss function for training dataset and validation dataset in the Y-Net training process, and it is clear that overfitting has not appeared. As illustrated in the enlarged portion of Fig. 7, LT and LX contribute almost equally to LTX and gradually converge, indicating that satisfactory reconstructions of temperature and H2O concentration can be simultaneously achieved.

 figure: Fig. 7.

Fig. 7. Evolution of the loss function for training dataset (total loss function of temperature and H2O concentration reconstruction (black solid line), loss function of temperature reconstruction (blue dashed dot line), and loss function of H2O concentration reconstruction (magenta dashed dot dot line), respectively), validation dataset (red dashed line) in the Y-Net training process.

Download Full Size | PDF

4.2 Evaluation of Y-Net

The complete training process takes typically 50 ∼ 60 mins for the training dataset with 16000 samples. To further evaluate the performance of this deep learning network, relative errors between the ground truth and reconstructed temperature (H2O concentration) are defined:

$${e_T} = \frac{{{{\left|\left|{\overrightarrow {{T_G}} - \overrightarrow {{T_R}} } \right|\right|}_1}}}{{{{\left|\left|{\overrightarrow {{T_G}} } \right|\right|}_1}}}$$
$${e_X} = \frac{{{{\left|\left|{\overrightarrow {{X_G}} - \overrightarrow {{X_R}} } \right|\right|}_1}}}{{{{\left|\left|{\overrightarrow {{X_G}} } \right|\right|}_1}}}$$
where ${\left|\left|{} \right|\right|_1}$ denotes the Manhattan norm.

Shown in Fig. 8 are three demonstrative cases with the phantoms and corresponding reconstruction errors in the absence of noise. A typical reconstruction time for the test dataset with 2000 samples is on the order of milliseconds using the proposed Y-Net. Reconstructions of temperature and concentration for the three cases with different multi-modal distributions exhibit satisfactory similarity (i.e., locations, sizes, magnitudes) with the ground truth. Much larger discrepancies are observed in the distributions of $\Delta T$ and $\Delta {X_{{H_2}O}}$ due to some low values on the edges of the sinograms, as illustrated in Fig. 3. The average reconstruction errors of temperature and H2O concentration for the testing dataset with 2000 samples are 1.55% and 2.47%, with standard deviations of 0.46% and 0.75%, respectively. Fig. 9 shows the NUC value distributions of the testing dataset with 2000 samples, together with the performance of tomographic reconstructions in the absence of noise with the proposed Y-Net under different NUC conditions. The error bars shown in Fig. 9 are calculated from the standard deviations of the reconstruction errors. It is apparent that the reconstruction errors for both temperature and species concentration distributions increase almost linearly with increasing NUC. The linear relationship may stem from the fact that the phantoms are generated as a combination of a set of Gaussian distributions. The results indicate that the well-designed Y-Net can effectively extract the feature maps from the sinograms and simultaneously achieve satisfactory reconstructions of temperature and H2O concentration within the NUC range of 0.02∼0.20. Assuming that the linear relationship holds between NUTX and eX, the proposed Y-Net still allows satisfactory reconstruction with eX equal to 5.0% of H2O concentration distribution at NUCX = 0.64.

 figure: Fig. 8.

Fig. 8. Three cases with representative phantoms and the corresponding reconstruction errors in the absence of noise. Case 1: (a) ∼ (d), Case 2: (e) ∼ (h), Case 3: (i) ∼ (l).

Download Full Size | PDF

 figure: Fig. 9.

Fig. 9. NUC value distributions of the testing dataset with 2000 samples, together with the performance of tomographic temperature (a) and species concentration (b) reconstructions with the proposed Y-Net under different NUC conditions.

Download Full Size | PDF

However, as discussed in [38], for nonlinear tomographic problem, the projections are nonlinear and sensitive to temperature, but linear and relatively insensitive to concentration. This further increase the complexity and difficulty in the reconstruction of concentration. In this work, Y-Net with a dual-branch symmetrical architecture is utilized for proof-of-concept demonstrations and more sophisticated neural network or fine-tuning of the hyper-parameter $\lambda (0 < \lambda < 1)$ could be implemented to further improve the reconstruction of H2O concentration.

Since the U-Net is tailored for applications with small datasets, the performance of the proposed Y-Net with different sizes of dataset needs to be further investigated. In addition, four projections, each with 24 parallel beams are assumed in this work, resulting in a total number of 96 LOS measurements projections. However, in practical combustion environments, the achievable number of projections can be limited. Therefore, the same reconstruction process is repeated and tested with smaller datasets and limited projections. The latter is implemented by increasing the distance between each parallel laser beam from d to 2d or using only two orthogonal projections $(\alpha \textrm{ = }{\textrm{0}^\circ },9{\textrm{0}^\circ }).$ More specifically, three sets of different optical configurations (two orthogonal projections each with 12 or 24 parallel beams, and four projections each with 12 parallel beams) are implemented. Accordingly, this will result in three different sizes of input matrices of $12 \times 12,12 \times 24,24 \times 12$ for the proposed Y-Net.

Fig. 10 (a) and (b) show the resulted reconstruction errors of temperature and H2O concentration for different numbers of dataset (2000, 12000, 20000) and optical configurations (sinograms with size of $12 \times 12,24 \times 12,12 \times 24,24 \times 24$). Each dataset is again split into 80%, 10% and 10% for training, validation and testing, respectively. Improved reconstructions for both temperature and H2O concentration are observed in Fig. 10 (a) and (b) with the number of datasets increasing. It is worth noting that even training the Y-Net with a dataset of as small as 2000 samples, the reconstruction errors for all four optical configurations are within 4.1 ∼ 5.4% and 6.0 ∼ 7.4% for temperature and H2O concentration, respectively. The performances of tomographic reconstruction based on the proposed Y-Net are on similar levels for the two configurations with sizes of $12 \times 24$ and $24 \times 24,$ indicating that satisfactory reconstructions can be achieved with only two orthogonal projections utilized. The proposed Y-Net works well for nonlinear tomographic absorption spectroscopy under small dataset and limited-projections conditions. However, such performance of the proposed Y-Net is closely related to the complexity, i.e., NUC value distributions, of the reconstructed fields, and further research is underway to reveal more deep-level relationship between the complexity and performance of tomographic problems other than the simple tendency already revealed here.

 figure: Fig. 10.

Fig. 10. Reconstruction errors of temperature (a) and H2O concentration (b) for various numbers of dataset and optical configurations.

Download Full Size | PDF

4.3 Comparison between Y-Net and SA

It has been verified previously that SA algorithm is capable of solving a nonlinear minimization problem in tomographic reconstruction problems [18]. To further investigate the advantages of the proposed Y-Net, comparison with previous state-of-the-art SA algorithm is demonstrated. Three representative cases shown in Fig. 8, with different multi-modal distributions (i.e., locations, sizes, magnitudes), are selected for demonstration. Random Gaussian noise with zero mean and standard deviations of 5%, 10% and 15%, are added to the sinograms of the three cases. The reconstruction errors of temperature and H2O concentration are estimated from 10 realizations at a given noise level for each reconstruction. SA takes about 15∼20 hours for each reconstruction, while a typical reconstruction time is on the order of milliseconds for the Y-Net, demonstrating its apparent superiority in computational efficiency. Take eT as an example, as seen from the upper panels of Fig. 11, the reconstruction errors for both Y-Net and SA are almost equal in the absence of noise. With the increase of noise level, the reconstruction errors of SA increase significantly, while Y-Net exhibits better noise immunity. The same stands true for H2O concentration reconstruction, as illustrated in the lower panels of Fig. 11. The uncertainties of the reconstruction errors for both Y-Net and SA, are estimated from 10 repeated realizations at each Gaussian noise level. The temperature reconstruction error of Y-Net stays below 3.7% even at 15% relative noise level. This is consistent with the results reported elsewhere [39,40] that deep-learning based methods have better noise suppression and removal capability.

 figure: Fig. 11.

Fig. 11. Comparison of reconstruction errors of temperature and H2O concentration between the proposed Y-Net and SA algorithm for three representative cases shown in Fig. 7. Case 1 (Left panels), Case 2 (Middle panels) and Case 3 (Right panels).

Download Full Size | PDF

Moreover, as discussed in [38], when solving a nonlinear minimization problem with SA, it is critical to determine the regularization parameters in order to incorporate the constraints for smoothness in tomographic reconstruction. Nevertheless, such a task is not always trivial to accomplish. In this work, the optimal regularization parameters are determined based on the two-step methods developed in [38]. However, it has to be noted that for the reconstruction tasks of different temperature and H2O concentration distributions, the optimal parameters may not always be the same and thus need to be determined separately. On the other hand, the proposed Y-Net has greater adaptability and robustness in tomographic reconstruction of various cases within the NUC range of 0.02∼0.20 compared to the SA method. More detailed comparisons between Y-Net and SA can be implemented for various temperature and concentration distributions, but it is out of the technical scope of the current study.

In summary, in addition to the inherited advantages of deep-learning based methods, such as reconstruction accuracy, noise immunity, and computational efficiency, the Y-Net enables simultaneous reconstruction of both temperature and species concentration. It suggests that tomographic absorption sensor with Y-Net architecture is a promising technique for real-time, in situ monitoring of practical combustion environments. Further work could be conducted on more advanced neural network design such as the incorporation of inception modules to simplify the network architecture [41], investigations between the complexity and performance of tomographic problems, and experimental demonstrations in practical flames. In addition, variants of Y-Net developed for multi-output tomographic problems have been reported [36]. For more complicated practical combustion environments with unknown temperature, species concentration and pressure distributions, more advanced neural network architecture can be conceivably devised such as incorporating another branch for pressure distribution reconstruction.

5. Conclusions

This work reports the first dual-branch Y-Net, incorporating calibration-free WMS with deep learning network, to solve nonlinear tomographic problems. Previous demonstration of multi-field reconstruction has only explored linear tomographic problems, and only preliminary study of nonlinear DAS-based tomographic deep learning methods has been demonstrated to reconstruct either temperature or species concentration alone, neglecting the inter-dependence of temperature and species concentration. Some key features of the proposed method are worth highlighting: (1) In addition to inheriting the advantages of deep learning methods such as better noise immunity, higher reconstruction accuracy and computational efficiency, the proposed Y-Net enables simultaneous reconstruction of both temperature and species concentration; (2) NUC method can serve as a benchmark to quantify the complexity (i.e., the non-uniformity) of the reconstructed temperature and species concentration fields in tomographic problems, and only minor performance deterioration is seen with increased field complexity. The results provide guidelines to further reveal more deep-level relationship between the complexity and performance of tomographic problems.

Specifically, a total number of 20000 samples are randomly generated, of which 80% are used for training, 10% for validation, and 10% for testing. Each temperature or H2O concentration phantom contains three randomly positioned Gaussian distributions. The ROI is discretized into $24 \times 24,$ pixels and 24 parallel beams are used at each projection angle $(\alpha \textrm{ = }{\textrm{0}^\circ },{45^\circ },9{\textrm{0}^\circ },{135^\circ }),$ resulting in a total number of 96 LOS measurements projections. A dual-branch Y-Net consisting of an encoder-decoder network with sum-skip-concatenation connections, is proposed with a total of 87 well-designed layers. Temperature and species concentration distributions can be simultaneously reconstructed with high accuracy, and the average reconstruction errors of temperature and H2O concentration are 1.55% and 2.47% for the testing dataset with 2000 samples. With NUC increasing from 0.02 to 0.20, the reconstruction errors for both temperature and species concentration distributions increase moderately and almost linearly. A relatively satisfactory reconstruction of H2O concentration distribution with eX equal to 5.0%, using the proposed Y-Net, can be achieved at NUCX = 0.64.

The performance of the proposed Y-Net under small dataset and limited-projection-number conditions is further investigated and the reconstruction errors for all four optical configurations with a dataset of as small as 2000 samples, are within 4.1 ∼ 5.4% and 6.0 ∼ 7.4% for temperature and H2O concentration, respectively. Moreover, Y-Net shows its superiority in noise immunity and computational efficiency in nonlinear tomographic problems over the state-of-the-art simulated annealing algorithm. In summary, a dual-branch deep learning network (Y-Net) has been demonstrated for WMS-based nonlinear tomography to simultaneously reconstruct temperature and species concentration fields. It is envisioned that such method can be an effective solution for achieving spatially-resolved, real-time, in situ monitoring in complicated practical combustion environments.

Funding

National Natural Science Foundation of China (51976105, 61627804, 91841302).

Disclosures

The authors declare no conflicts of interest.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. R. K. Hanson, “Applications of quantitative laser sensors to kinetics, propulsion and practical energy systems,” Proc. Combust. Inst. 33(1), 1–40 (2011). [CrossRef]  

2. C. S. Goldenstein, R. M. Spearrin, J. B. Jeffries, and R. K. Hanson, “Infrared laser-absorption sensing for combustion gases,” Prog. Energy Combust. Sci. 60, 132–176 (2017). [CrossRef]  

3. Z. Wang, P. Fu, and X. Chao, “Laser absorption sensing systems: challenges, modeling, and design optimization,” Appl. Sci. 9(13), 2723 (2019). [CrossRef]  

4. K. Sun, X. Chao, R. Sur, C.S. Goldenstein, J. B. Jeffries, and R. K. Hanson, “Analysis of calibration-free wavelength-scanned wavelength modulation spectroscopy for practical gas sensing using tunable diode lasers,” Meas. Sci. Technol. 24(12), 125203 (2013). [CrossRef]  

5. C.S. Goldenstein, C. S. Strand, I. A. Schultz, K. Sun, J. B. Jeffriess, and R. K. Hansons, “Fitting of calibration-free scanned-wavelength-modulation spectroscopy spectra for determination of gas properties and absorption lineshapes,” Appl. Opt. 53(3), 356–367 (2014). [CrossRef]  

6. S. T. Sanders, J. Wang, J. B. Jeffries, and R. K. Hanson, “Diode-Laser Absorption Sensor for Line-of-Sight Gas Temperature Distributions,” Appl. Opt. 40(24), 4404–4415 (2001). [CrossRef]  

7. X. Liu, J. B. Jeffries, and R. K. Hanson, “Measurement of Non-Uniform Temperature Distributions Using Line-of-Sight Absorption Spectroscopy,” AIAA J. 45(2), 411–419 (2007). [CrossRef]  

8. C. W. Cai and F. Kaminski, “Tomographic absorption spectroscopy for the study of gas dynamics and reactive flows,” Prog. Energy Combust. Sci. 59, 1–31 (2017). [CrossRef]  

9. C. Liu and L. Xu, “Laser absorption spectroscopy for combustion diagnosis in reactive flows: A review,” Appl. Spectrosc. Rev. 54(1), 1–44 (2019). [CrossRef]  

10. C. Wei, D. I. Pineda, C. S. Goldenstein, and R. M. Spearrin, “Tomographic laser absorption imaging of combustion species and temperature in the mid-wave infrared,” Opt. Express 26(16), 20944–20951 (2018). [CrossRef]  

11. C. Wei, K. K. Schwarm, D. I. Pineda, and R. M. Spearrin, “Volumetric laser absorption imaging of temperatu-re, CO and CO2 in laminar flames using 3D masked Tikhonov regularization,” Combust. Flame 224, 239–247 (2021). [CrossRef]  

12. C. Wei, K. K. Schwarm, D. I. Pineda, and R. M. Spearrin, “Physics-trained neural network for sparse-view volumetric laser absorption imaging of species and temperature in reacting flows,” Opt. Express 29(14), 22553–22566 (2021). [CrossRef]  

13. S. R. Arridge and J. C. Schotland, “Optical tomography: forward and inverse problems,” Inverese. Probl. 25(12), 123010 (2009). [CrossRef]  

14. S. J. Grauer and A. M. Steinberg, “Linear absorption tomography with velocimetry (LATV) for multiparameter measurements in high-speed flows,” Opt. Express 28(22), 32676–32692 (2020). [CrossRef]  

15. S. J. Grauer, J. Emmert, S. T. Sanders, S. Wagner, and K. J. Daun, “Multiparameter gas sensing with linear hyperspectral absorption tomography,” Meas. Sci. Technol. 30(10), 105401 (2019). [CrossRef]  

16. Y. Jiang, J. Si, R. Zhang, G. Enemali, B. Zhou, H. McCann, and C. Liu, “CSTNet: A Dual-Branch Convolutio-nal Network for Imaging of Reactive Flows using Chemical Species Tomography,” arXiv:2010.03868 (2020).

17. L. Ma, W. Cai, A. W. Caswell, T. Kraetschmer, S. T. Sanders, S. Roy, and J. R. Gord, “Tomographic imaging of temperature and chemical species based on hyperspectral absorption spectroscopy,” Opt. Express 17(10), 8602–8613 (2009). [CrossRef]  

18. W. Cai and C. F. Kaminski, “Multiplexed absorption tomographyss with calibration-free wavelength modulation spectroscopy,” Appl. Phys. Lett. 104(15), 154106 (2014). [CrossRef]  

19. J. Huang, H. Liu, J Dai, and W. Cai, “Reconstruction for limited-data nonlinear tomographic absorption spectroscopy via deep learning,” J. Quant. Spectrosc. Radiat. Transfer 218, 187–193 (2018). [CrossRef]  

20. A. Deng, J. Huang, H. Liu, and W. Cai, “Deep learning algorithms for temperature field reconstruction of nonlinear tomographic absorption spectroscopy,” Meas: Sens. 10, 100024 (2020). [CrossRef]  

21. J. Shi, H. Qi, J. Zhang, Y. Ren, L. Ruan, and Y. Zhang, “Simultaneous measurement of flame temperature and species concentration distribution from nonlinear tomographic absorption spectroscopy,” J. Quant. Spectrosc. Radiat. Transfer 241, 106693 (2020). [CrossRef]  

22. G. Wang, J. C. Ye, K. Mueller, and J. A. Fessler, “Image reconstruction is a new frontier of machine learning,” IEEE Trans. Med. Imaging 37(6), 1289–1296 (2018). [CrossRef]  

23. G. Wang, J. C. Ye, and Man. B. De, “Deep learning for tomographic image reconstruction. Nature Machine Intelligence,” Nat Mach Intell 2(12), 737–748 (2020). [CrossRef]  

24. Q. Qu, Z. Cao, L. Xu, and W. Cai, “Optimal selection of spectral lines for multispectral absorption tomography,” Appl. Phys. B 124(9), 1–9 (2018). [CrossRef]  

25. I. E. Gordon, L. S. Rothman, R. J. Hargreaves, R. Hashemi, E. V. Karlovets, F. M. Skinner, E. K. Conway, C. Hill, R. V. Kochanov, Y. Tan, P. Wcisło, A. A. Finenko, K. Nelson, P. F. Bernath, M. Birk, V. Boudon, A. Campargue, K. V. Chance, A. Coustenis, B. J. Drouin, J. –M. Flaud, R. R. Gamache, J. T. Hodges, D. Jacquemart, E. J. Mlawer, A. V. Nikitin, V. I. Perevalov, M. Rotger, J. Tennyson, G. C. Toon, H. Tran, V. G. Tyuterev, E. M. Adkins, A. Baker, A. Barbe, E. Canè, A. G. Császár, A. Dudaryonok, O. Egorov, A. J. Fleisher, H. Fleurbaey, A. Foltynowicz, T. Furtenbacher, J. J. Harrison, J. –M. Hartmann, V. –M. Horneman, X. Huang, T. Karman, J. Karns, S. Kassi, I. Kleiner, V. Kofman, F. Kwabia-Tchana, N. N. Lavrentieva, T. J. Lee, D. A. Long, A. A. Lukashevskaya, O. M. Lyulin, V. Yu. Makhnev, W. Matt, S. T. Massie, M. Melosso, S. N. Mikhailenko, D. Mondelain, H. S. P. Müller, O. V. Naumenko, A. Perrin, O. L. Polyansky, E. Raddaoui, P. L. Raston, Z. D. Reed, M. Rey, C. Richard, R. Tóbiás, I. Sadiek, D. W. Schwenke, E. Starikova, K. Sung, F. Tamassia, S. A. Tashkun, J. Vander. Auwera, I. A. Vasilenko, A. A. Vigasin, G. L. Villanueva, B. Vispoel, G. Wagner, A. Yachmenev, and S. N. Yurchenko, “The HITRAN 2020 molecular spectroscopic database,” J. Quant. Spectrosc. Radiat. Transfer 107949, 1 (2021). [CrossRef]  

26. Q. Xiao, Y. Zhai, Z. Lv, J. Xu, J. Pan, and H. Wang, “Non-uniformity quantification of temperature and concentration fields by statistical measure and image analysis,” Appl. Therm. Eng. 124, 1134–1141 (2017). [CrossRef]  

27. G. Rodriguez, T. Anderlei, M. Micheletti, M. Yianneskis, and A. Ducci, “On the measurement and scaling of mixing time in orbitally shaken bioreactors,” Biochem. Eng. J. 82, 10–21 (2014). [CrossRef]  

28. A. Amdouni, P. Castagliola, H. Taleb, and G. Celano, “One-sided run rules control charts for monitoring the coefficient of variation in short production runs,” Eur. J. Ind. Eng. 10(5), 639–663 (2016). [CrossRef]  

29. G. Jumarie, “Entropy of markovian processes: Application to image entropy in computer vision,” J. Franklin Inst. 335(7), 1327–1338 (1998). [CrossRef]  

30. O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention (2015), pp. 234–241), Springer, Cham.

31. Y. Han and J. C. Ye, “Framing U-Net via deep convolutional framelets: Application to sparse-view CT,” IEEE Trans. Med. Imaging 37(6), 1418–1429 (2018). [CrossRef]  

32. J. C. Ye, Y. Han, and E. Cha, “Deep convolutional framelets: A general deep learning framework for inverse problems,” SIAM J. Imaging Sci. 11(2), 991–1048 (2018). [CrossRef]  

33. A. Mohammed, S. Yildirim, I. Farup, M. Pedersen, and Ø. Hovde, “Y-net: A deep convolutional neural network for polyp detection,” arXiv:1806.01907 (2018).

34. H. Lan, D. Jiang, C. Yang, and F. Gao, “Y-Net: a hybrid deep learning reconstruction framework for photoacoustic imaging in vivo,” arXiv:1908.00975 (2019).

35. K. Wang, J. Dou, Q. Kemao, J. Di, and J. Zhao, “Y-Net: a one-to-two deep learning framework for digital holographic reconstruction,” Opt. Lett. 44(19), 4765–4768 (2019). [CrossRef]  

36. K. Wang, Q. Kemao, J. Di, and J. Zhao, “Y4-Net: a deep learning solution to one-shot dual-wavelength digital holographic reconstruction,” Opt. Lett. 45(15), 4220–4223 (2020). [CrossRef]  

37. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv:1412.6980 (2014).

38. L. Ma and W. Cai, “Determination of the optimal regularization parameters in hyperspectral tomography,” Appl. Opt. 47(23), 4186–4192 (2008). [CrossRef]  

39. C. Tian, L. Fei, W. Zheng, Y. Xu, W. Zuo, and C. W. Lin, “Deep learning on image denoising: An overview,” Neural. Netw. 131, 251–275 (2020). [CrossRef]  

40. M. Aghajarian, J. E. McInroy, and S. Muknahallipatna, “Deep learning algorithm for Gaussian noise removal from images,” J. Electron. Imag. 29(4), 043005 (2020). [CrossRef]  

41. C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit.2818–2826 (2016).

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (11)

Fig. 1.
Fig. 1. Schematic diagram of multiplexed nonlinear absorption tomography based on calibration-free wavelength modulation spectroscopy.
Fig. 2.
Fig. 2. Calculated WMS-2f/1f signal with different temperature and H2O concentration, respectively.
Fig. 3.
Fig. 3. Two examples of representative phantoms with different multi-modal distributions of temperature and H2O concentration, and the corresponding WMS-2f/1f. Example 1: (a) ∼ (c), Example 2: (d) ∼ (f). The red crosses are at the center of the phantoms. Horizontal axes of Fig. (c) and (f) correspond to 24 LOS measurement projections at each projection angles, i.e., $(\alpha \textrm{ = }{\textrm{0}^\circ },{45^\circ },9{\textrm{0}^\circ },{135^\circ }),$ respectively.
Fig. 4.
Fig. 4. Simplified U-Net architecture.
Fig. 5.
Fig. 5. Global architecture of Y-Net in this work.
Fig. 6.
Fig. 6. NUC value distributions of 20000 temperature (a) and species concentration (b) phantoms, respectively.
Fig. 7.
Fig. 7. Evolution of the loss function for training dataset (total loss function of temperature and H2O concentration reconstruction (black solid line), loss function of temperature reconstruction (blue dashed dot line), and loss function of H2O concentration reconstruction (magenta dashed dot dot line), respectively), validation dataset (red dashed line) in the Y-Net training process.
Fig. 8.
Fig. 8. Three cases with representative phantoms and the corresponding reconstruction errors in the absence of noise. Case 1: (a) ∼ (d), Case 2: (e) ∼ (h), Case 3: (i) ∼ (l).
Fig. 9.
Fig. 9. NUC value distributions of the testing dataset with 2000 samples, together with the performance of tomographic temperature (a) and species concentration (b) reconstructions with the proposed Y-Net under different NUC conditions.
Fig. 10.
Fig. 10. Reconstruction errors of temperature (a) and H2O concentration (b) for various numbers of dataset and optical configurations.
Fig. 11.
Fig. 11. Comparison of reconstruction errors of temperature and H2O concentration between the proposed Y-Net and SA algorithm for three representative cases shown in Fig. 7. Case 1 (Left panels), Case 2 (Middle panels) and Case 3 (Right panels).

Tables (1)

Tables Icon

Table 1. Spectroscopic data for the H2O transitions of interest taken from HITRAN 2020

Equations (19)

Equations on this page are rendered with MathJax. Learn more.

v ( t ) = v ¯ + a m cos ( 2 π f m t )
I 0 ( t ) = I 0 ¯ [ 1 + i 0 cos ( 2 π f m t + φ 1 ) + i 2 cos ( 4 π f m t + φ 2 ) ]
I t ( t )  =  G I 0 ( t ) τ [ v ( t ) ] = G I 0 ( t ) k = 0 H k ( v , a m ) cos ( 2 π f m t )
S 1 f  =  G I 0 ¯ 2 { [ H 1 + i 0 ( H 0 + H 2 2 ) cos ( φ 1 ) + i 2 2 ( H 1 + H 3 ) cos ( φ 2 ) ] 2 + [ i 0 ( H 0 H 2 2 ) sin ( φ 1 ) + i 2 2 ( H 1 H 3 ) sin ( φ 2 ) ] 2 } 1 / 2
S 2 f  =  G I 0 ¯ 2 { [ H 2 + i 0 2 ( H 1 + H 3 ) cos ( φ 1 ) + i 2 ( H 0 + H 4 2 ) cos ( φ 2 ) ] 2 + [ i 0 2 ( H 1 H 3 ) sin ( φ 1 ) + i 2 ( H 0 H 4 2 ) sin ( φ 2 ) ] 2 } 1 / 2
[ ( S 2 f / 1 f ) v ] L × M  =  [ ( S 2 f / 1 f ) v 1 , 1 ( S 2 f / 1 f ) v 1 , 2 ( S 2 f / 1 f ) v 1 , M ( S 2 f / 1 f ) v 2 , 1 ( S 2 f / 1 f ) v 2 , 2 ( S 2 f / 1 f ) v 1 , M ( S 2 f / 1 f ) v L , 1 ( S 2 f / 1 f ) v L , 2 ( S 2 f / 1 f ) v L , M ] L × M
T ( x , y )  =  T min + 1 3 i = 1 3 ξ i ( T max T min ) exp [ ( x x c , i ) 2 + ( y y c , i ) 2 2 σ T 2 ]
X ( x , y )  =  X min + 1 3 i = 1 3 ξ i ( X max X min ) exp [ ( x x c , i ) 2 + ( y y c , i ) 2 2 σ X 2 ]
d P Q ( θ )  =  j = 1 θ 2 i = 1 θ 1 T i , j j = 1 D i = 1 D T i , j  -  θ 1 θ 2 D 2
N U C q T = sup | d P Q ( θ , q ) |
N U C q T = max { N U C 1 T , N U C 2 T , N U C 3 T , N U C 4 T }
I d = φ ( I in ϕ ¯ )
φ  =  [ b 1 1 b 1 0 0 0 0 0 0 0 0 b n 2 1 b n 2 ] R n 2 × n
I out = I in ϕ ¯ + φ T φ ( I in ϕ ¯ )
L T X = λ L T + L X
L T = 1 N n N | | T n G T n R | | 2
L X = 1 N n N | | X n G X n R | | 2
e T = | | T G T R | | 1 | | T G | | 1
e X = | | X G X R | | 1 | | X G | | 1
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.