Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Deep-learning-enhanced ice thickness measurement using Raman scattering

Open Access Open Access

Abstract

In ice thickness measurement (ICM) procedures based on Raman scattering, a key issue is the detection of ice–water interface using the slight difference between the Raman spectra of ice and water. To tackle this issue, we developed a new deep residual network (DRN) to cast this detection as an identification problem. Thus, the interface detection is converted to the prediction of the Raman spectra of ice and water. We enabled this process by designing a powerful DRN that was trained by a set of Raman spectral data, obtained in advance. In contrast to the state-of-the-art Gaussian fitting method (GFM), the proposed DRN enables ICM with a simple operation and low costs, as well as high accuracy and speed. Experimental results were collected to demonstrate the feasibility and effectiveness of the proposed DRN.

© 2019 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

The polar regions have been receiving intense scientific and technological interest, due to global climate change and increasing industrial activity. In the polar regions, ice acts as an effective intermediary between the ocean and the atmosphere by modulating heat, mass, and momentum transfer [1]. As a result, ice studies help researchers to model, analyze, and forecast global climate change. As an important indicator, ice thickness measurements (ICMs) play an important role in ice studies. In addition, ICM can also be used to ensure safe voyages and explore resources [24].

Various ICM approaches have been developed, with varying degrees of accuracy, spatial and temporal resolution, capability, and cost. Drilling is the most common method and has the benefits of a simple procedure and accuracies on the scale of millimeters [5,6]. However, these benefits come at the cost of time, manpower, and requirement of working on the ice owing to its single-point measurement procedure. Thermistors can be employed in long-term range ICM; however, this method is seldom adopted in practice, as it requires massive manpower to recover equipment from the ice, and, in particular, it can only marginally distinguish the interface between melting ice and water, owing to the similar temperatures they share. Electro-magnetic sounding provides an excellent sampling rate [6]; however, its accuracy relies heavily upon ice salinity, porosity, and further calibrations. Upward-looking sonars are costly and commonly available for long-term measurements [7]. However, the ice velocity drift must also be obtained simultaneously to resolve the measurement correctly. Satellite remote sensing offers repeat data over large areas [8,9]. However, its accuracy is low and strongly influenced by weather conditions. Recently, a compact Raman LIDAR system with a spectrograph was developed to enable ICM [10,11]. The difference between the Raman spectra of ice and liquid water was used to identify the ice–water interface, while the elastic signal was employed to determine the air–ice interface. This system provides a simple implementation and does not require contact with the ice in its setup. Hence, it can be installed on an unmanned airplane or autonomous underwater vehicle. However, it requires sequential Raman data not only of the ice but also of the water to resolve the ice-water interface. This results in a high workload.

As one of the most attractive research fields in recent years, deep learning allows complex operations on raw datasets using multilayered artificial neural networks [1214]. It has shown remarkable success in signal-extracting, feature-learning, and modelling complex relationships. However, deep learning for Raman studies remains a comparatively unexplored area. Liu et al. [15] proposed a unified solution to identify chemical species via deep learning, in which a convolutional neural network was trained to automatically identify substances according to their Raman spectrum, without preprocessing procedures. Krauß et al. [16] applied deep learning to the whole-image classifiers for Raman-microscopy-based cytopathology, thus resulting in a high accuracy. Fan et al. [17] proposed an approach titled “deep learning-based component identification for the Raman spectra of mixtures”. To the best of our knowledge, there is no existing work demonstrating deep learning approaches to deal with raw Raman spectra for ICM.

In this paper, we suggest an improved deep residual network (DRN) for ICM, using Raman scattering to facilitate a substantial performance enhancement. In our approach, the DRN is employed to extract the ice-water border from the Raman spectra. Experimental results demonstrate that, in contrast to [10,11], our approach not only greatly simplifies the work load at low costs, but also enables higher accuracy and faster operation times.

2. Principle

2.1 Experimental setup

Figure 1 shows the experimental setup, the sample is excited by an argon-ion laser at 514 nm. The output beam is collimated and expanded by a beam expander BE, then converged by an objective lens (OL) with a focal length of 180 mm, to illuminate the remote sample. The beam scattered from the sample is collected by OL again and then guided via a beam splitter BS to converge onto a spectrometer (AvaSpec-HS1024*122TEC with resolution of 1.5 nm and spectral range of 200–1160 nm) via a collection lens CL. An ice cube was formed from tap water in a laboratory refrigerator, it had a good optical quality with a few defects such as bubbles, brines, and cracks inside. The ice sample was then put into a jar, where it floated on top of tap water. During the measurement process, we varied the lens-to-sample distance and detected signals from different laser beam waist positions. In this scenario, the focal length of OL should be greater than the ice thickness but as small as possible to optimize the spatial resolution. Thus, we affix the lens OL onto a micro-motional scanning stage, to search the air–ice interface and the ice–water interface, respectively. The air–ice interface can be estimated at the maximum value of the elastic signal, whose position is set to zero in the following analysis. However, as pointed out by [10], the identification of the ice–water interface is a key issue, in which the Raman OH-band profiles need to be calculated. In [11], the profile was fitted with a Gaussian fitting method (GFM), by plotting the corresponding curve centers as functions of the lens-to-sample distance. However, this method required a series of Raman spectra from ice and water, resulting in an extremely complex measuring process with low speed. To simplify the process and ensure high speed at high accuracies, we develop an improved DRN to recognize the Raman signal. To investigate the generalization of the improved DRN in spectral recognition, we carry out the following steps to obtain datasets, and conduct classification experiments. First, after an ice cube is inserted into the water jar, multiple Raman signals are acquired along the optical axis from ice to water using the apparatus shown in Fig. 1. To make the data more representative, we repeat the data acquisition in random positions with different sizes of ice cubes. As a result, we obtain a large number of Raman signals for both floating ice and water. Second, the Raman signals are divided into two types according to the position of the laser focal point. In such cases, the one hot code [18] is employed to represent the Raman spectra of ice and water, with (0, 1) as ice and (1, 0) as water, in which the ratio between them is about 1:1. Third, after collecting signals of multiple ice cubes, we obtain 2290 Raman datasets. Finally, we assess the model through the classic train-validation-test approach. During the assessment process, the aforementioned Raman datasets are randomly separated into two subsets, with a ratio of 85:15 for training and validation, respectively, while other different Raman datasets are used for testing. To relieve the effect of the differences of the spectral peak intensities on the measuring accuracy, and to speed up the learning convergence, all the Raman spectra data used in the experiments are normalized.

 figure: Fig. 1.

Fig. 1. Experimental setup for ICM: M1 and M2 represent the mirrors; BE the beam expander; BS the beam splitter; OL the objective lens; and CL the collection lens.

Download Full Size | PDF

2.2 Improved DRN

Deep learning has been shown to be effective for a variety of recognition, classification, and segmentation tasks [1926]. Normally, deep learning models share convolution kernels to exploit the local features of data and then, extract global training features. In the traditional convolutional neural network (CNN) [2023], as the number of network layers deepens, the problem of gradient disappearance/explosion arises, resulting in a decrease in accuracy. DRNs [2729] were proposed to solve this issue. In DRNs, the input information is directly connected to the output, to protect the integrity of the information. Unlike traditional networks, which are the DRNs used for image classification, we develop an improved DRN for one-dimensional (1D) Raman spectral recognition.

The basic architecture of the improved DRN is shown in Fig. 2. There are several functional layers, including residual convolutional layer, pooling layer, fully connected layer, and output layer. Assuming that for the k-th convolutional layer, the number of the convolution kernel is K(k), and the convolution result is cjk (j = 1, 2, …, K(k)). The convolutional layer in each block can be described as

$$c_j^{(k )} = R\left[ {\sum\limits_{i = 1}^{{K^{({k - 1} )}}} {c_i^{({k - 1} )} \ast \omega_{ij}^{(k )} + b_j^{(k )}} } \right]$$
where ωij is the weight and bj is the bias, both of which are learned through training, R(.) represents an activation function, and * denotes convolution. In our work, the rectified linear unit (ReLU) of R(x) = max (0, x) is employed to settle the problem of nonlinear classification. To improve the network performance, a regularization coefficient [30] is used to constrain the convolutional layer and prevent the network from over-fitting.

 figure: Fig. 2.

Fig. 2. Basic architecture of the improved DRN. In each layer, a residual convolutional layer and a max pooling are involved, in which “×3” on the top of “Block” indicates repeating the block three times. “FC1” and “FC2” denote fully connected layers, “Dropout” represents the dropout layer. In the Feature Extraction, the kernel size and depth are demonstrated at the bottom of each layer, in which “depth_n” (n = 1, 2, …, 7) means 32, 32, 64, 64, 128, 128, and 256, respectively. In the Regression, the input Raman signals are predicted with a response, denoting the estimated position of the ice–water interface.

Download Full Size | PDF

After each application of the ReLU, the intermediate data is divided equally and then, added to enable information fusion, which helps to enhance the extraction capacity of the underlying characterization. The combination of convolution and fusion is the essential building block of our network and is repeated three times. To ease the training of networks and gain accuracy on the feature extraction of the spectral peak, a shortcut flow is introduced to complete the residual convolutional layer. This shortcut flow is marked by the blue line [2729]. We define the input data as x and the output data as y. The residual is F(x), and the residual module can then be written as

$$y = F(x )+ x$$
Then the residual convolutional layer is followed by a maximum pooling layer [18,22] to down-sample the feature map but retain the main features as much as possible, the schematic for the maximum pooling is shown in Fig. 3. As shown in Fig. 3, we employ the maximum value as the representation over a small region. This operation greatly reduces the spatial dimension of the representation and the number of parameters, subsequently reducing the computation cost. It could also help the network to prevent over-fitting. The residual convolution and maximum pooling form a complex layer which is repeated seven times, thus resulting in the final feature extraction module, as shown in Fig. 2. This module is employed to identify the implicit features of the Raman spectra.

 figure: Fig. 3.

Fig. 3. Framework of the max pooling.

Download Full Size | PDF

Thereafter, the extracted feature information is sent into the fully connected layer (FCL) [18]. To avoid over-fitting while increasing the learning speed, a dropout layer [31] is introduced to guard the network. Through each training iteration, the individual nodes are either retained (with a probability of p) or cast away from the network (with a probability of 1−p), thereby yielding a compressed network [31]. The combination of the FCL and dropout layer is grouped as the regression module for regression analysis, which is repeated twice. Finally, the output layer gives a predicted response.

To train the network, we need a substantial number of Raman signals, including those of ice and water. Thereafter, we assign the real code as the response to each Raman signal, as shown in Section 2.1. Suppose y is the estimated output and yr is the corresponding real code. In such a case, the one-hot code of yr indicates the probability distribution under ideal conditions, in which (1, 0) represents water, and (0, 1) represents ice. As a result, each neuron array in the output layer becomes a 1 × 2 array. The network training is completed by minimizing the loss function. In general, the loss function L(·) [32] for each sample is a multi-class cross entropy loss function, which can be described as follows:

$$L (y,{y_r}) ={-} \frac{1}{N}\left\{ {\sum\limits_{i = 1}^N {{y_{r\_i}} \times \log [{{\mathop{\textrm {softmax}}\nolimits} ({y_i})} ]} } \right\}$$
where N is the number of categories, softmax(·) represents the possibility of the model identifying the sample as the i-th category, which can be written as
$${\mathop{\textrm {softmax}}\nolimits} ({y_i}) = \frac{{e_{}^{{y_i}}}}{{\sum\limits_{\textrm{j} = 1}^N {e_{}^{{y_j}}} }}$$
In the proposed DRN, to avoid the overfitting and encourage its generalization, the L2 regularization [30] is introduced to restrict the loss function as follows.
$$Loss = \frac{1}{M}\left[ {\sum\limits_{j = 1}^M {L({y_{}^j,y_r^j} )} } \right] + \alpha ||W ||_2^2$$
where M is the total number of mini-batches during the training, W is the weight vector, and α is a coefficient.

3. Results

To demonstrate the feasibility of the proposed DRN, we froze tap water in a laboratory refrigerator to form different sizes of ice cubes. The system shown in Fig. 1 was employed to collect Raman data. For comparison, the Raman spectra of water and ice were separately collected before the ice cube was inserted into the water jar, and the results are shown in Fig. 4(a). It is clear that there is a difference between the Raman spectra of water and ice before the insertion, which agrees well with the results obtained in [10] and [33]. Afterwards, the spectra collections were performed after the ice cube had been inserted into the water jar, and the results are shown in Fig. 4(b). It is notably difficult to distinguish the spectral differences between ice and water near the ice-water interface through manual classification alone. To search for an optimal method of performing data classification, we trained the proposed DRN to predict the two types of aforementioned samples. Meanwhile, we employed the one-hot code to identify the Raman spectra of ice and water.

 figure: Fig. 4.

Fig. 4. Typical Raman spectra of water and floating ice. (a) before ice insertion, (b) after ice insertion, in which “+” indicates the ice spectrum obtained above the ice-water interface, and “−“ indicates the water spectrum obtained beneath the ice-water interface.

Download Full Size | PDF

After acquiring 2290 Raman data with a ratio of 1:1 for ice and water, we randomly divided the data into two subsets, with a ratio of 85:15 for training and validation, respectively. The DRN is trained using the Adam optimizer [22], which is a form of gradient descent method. In this case, the learning rate was empirically set to 10−6 and allowed to decay with a rate of 0.99 at each training iteration. To solve the issues of (i) memory constraints and (ii) stagnation in local minima during optimization, at each iteration a single mini-batch [18,22] of 20 from the whole training data was sent into the network for training, instead of the entire set of Raman data. All the weights are initialized using truncated normalization with a standard deviation of 0.1, and the biases were initialized with a constant value of 0. In each mini-batch training step, one iteration of the optimization was conducted, and the parameters of the network were updated. The dropout probability was fixed at 0.5 during training, while in validation and testing it was set as 1. This enabled 50% of the nodes to be randomly selected and intentionally invalidated in training, to reduce over-fitting. The training environment was an Intel(R) Core(TM) i3–8100 CPU @3.60 GHz, with 8 GB memory.

Figure 5 shows the losses during the training process, in which the loss was recorded every fifteen iterations. It is clear that the proposed DRN converges gradually, since the loss descends along the course of training. After training for 7365 steps, the loss approached the order of 0.4. This accords with our expectation that the network is continuously updating its parameters and studying representative features of the Raman signals, resulting in a satisfactory training speed.

 figure: Fig. 5.

Fig. 5. Training loss during the training process.

Download Full Size | PDF

After training and validation of the DRN, a series of new Raman spectra, which had not appeared before, were fed into the network for testing. In this case, the Raman signals were acquired from different positions by varying the lens-to-sample distance with an increment of 1 mm. During testing, we not only ran the proposed network, but also compared it with the GFM. The results are illustrated in Fig. 6. To obtain a better comparison, the results are demonstrated side by side. For the GFM [10,11], the Raman OH-band profile was fitted by a Gaussian function, and the resulting curve centers were plotted as a function of the lens-to-sample distance. In this case, the center of the confidence band was taken as a quantitative indicator for the identification of the ice-water interface. The result is shown in Fig. 6(a). However, a group of Raman data near the ice-water interface (indicated by the green arrow line in Fig. 6(a)) required fitting with linear regression, thereby resulting in an extremely heavy measurement burden, and further limiting the measurement speed. Meanwhile, the measurement accuracy was determined by the confidence band interval around the inflection point. Therefore, apart from the micro-motional scanning stage, GFM also requires a high precision spectrometer, i.e., a spectrograph (Spectra Physics, MS127i with a resolution of 0.1 nm and spectral range of 500–750 nm), combined with a gated detector (ICCD, Andor iStar), to ensure measurement accuracy. However, in contrast to the GFM, the proposed DRN needs only a few data from near the ice-water interface (indicated by the green arrow line in Fig. 6(b)) and can immediately identify the ice-water interface once the water spectrum is determined. In other words, the proposed DRN requires neither the high precision spectrometer, nor a set of Raman data for the water from below the ice-water interface. After identifying the ice-water interface, the ICM can be completed by computing the distance between the air–ice interface and the ice–water interface. Thus, in contrast to the GFM, the proposed network can perform the ICM using a simple operation and at a low cost, whilst maintaining high accuracy and speed.

 figure: Fig. 6.

Fig. 6. ICM using (a) GFM and (b) the proposed DRN

Download Full Size | PDF

To further evaluate the performance, we compared the measured results of the GFM and proposed DRN using different thicknesses of ice, in which the nominal thickness of the ice is supplied by a Vernier caliper. The results are shown in Table 1, in which both absolute errors and relative errors have been calculated. It is obvious that in contrast to the GFM, the proposed DRN has a lower error. This further proves that the proposed DRN can provide high accuracy. It is noted that owing to the limitations imposed upon us by our ordinary laboratory conditions, such as the low laser power and small travel range of the scanning stage, the samples used for this study are less than 100.0 mm. To measure realistic samples (i.e., very thick samples), a Raman LIDAR with high power pulse laser [10,11] should be employed, as this would simultaneously measure the Raman spectra at multiple distances [34], or work in a single-photon counting and time-of-light measurement mode [35]. This will be a part of future study.

Tables Icon

Table 1. Comparison of the measurement error between DRN and GFM

4. Conclusion

In conclusion, we report an improved DRN to enable ICM based on Raman scattering. Unlike the conventional GFM, we consider the detection of the ice–water interface as an identification problem in which the DRN is employed to predict the Raman spectra of ice and water. The experimental results demonstrate that DRN can perform ICM with ease of operability and low cost, whilst maintaining high accuracy and speeds. We hope that the proposed method can be found to be practical in ICM applications.

Funding

National Natural Science Foundation of China (61775046); Natural Science Foundation of Heilongjiang Province (LC2018027).

Disclosures

The authors declare that there are no conflicts of interest related to this article.

References

1. V. V. Ivanov, V. A. Alexeev, I. Repina, N. V. Koldunov, and A. Smirnov, “Tracing Atlantic Water Signature in the Arctic Sea Ice Cover East of Svalbard,” Adv. Meteorol. 2012, 1–11 (2012). [CrossRef]  

2. A.-S. Milaković, B. Gunnarsson, S. Balmasov, S. Hong, K. Kim, P. Schütz, and S. Ehlers, “Current status and future operational models for transit shipping along the Northern Sea Route,” Mar. Policy 94, 53–60 (2018). [CrossRef]  

3. R. T. Datta, M. Tedesco, C. Agosta, X. Fettweis, P. K. Munneke, and M. R. van den Broeke, “Melting over the northeast Antarctic Peninsula (1999–2009): evaluation of a high-resolution regional climate model,” Cryosphere 12(9), 2901–2922 (2018). [CrossRef]  

4. A. F. Bunkin, V. K. Klinkov, V. N. Lednev, D. L. Lushnikov, A. V. Marchenko, E. G. Morozov, S. M. Pershin, and R. N. Yulmetov, “Remote sensing of seawater and drifting ice in Svalbard fjords by compact Raman lidar,” Appl. Opt. 51(22), 5477–5485 (2012). [CrossRef]  

5. K. V. Hoyland, “Ice thickness, growth and salinity in Van Mijenfjorden, Svalbard, Norway,” Polar Res. 28(3), 339–352 (2009). [CrossRef]  

6. K. Shirasawa, H. Eicken, K. Tateyama, T. Takatsuka, and T. Kawamura, “Sea-ice-thickness variability in the Chukchi Sea, spring and summer 2002–2004,” Deep Sea Res., Part II 56(17), 1182–1200 (2009). [CrossRef]  

7. A. Behrendt, W. Dierking, and H. Witte, “Thermodynamic sea ice growth in the central Weddell Sea, observed in upward-looking sonar data,” J. Geophys. Res. Oceans 120(3), 2270–2286 (2015). [CrossRef]  

8. R. Lindsay and A. Schweiger, “Arctic sea ice thickness loss determined using subsurface, aircraft, and satellite observations,” Cryosphere 9(1), 269–283 (2015). [CrossRef]  

9. N. S. Lewis, L. Koenig, G. Grant, D. Gallaher, K. Schaefer, J. Thompson, and G. G. Campbell, “Sea ice detection from persistent single-channel shortwave infrared satellite data,” Ecol. Inform. 52, 139–149 (2019). [CrossRef]  

10. S. M. Pershin, V. N. Lednev, V. K. Klinkov, R. N. Yulmetov, and A. F. Bunkin, “Ice thickness measurements by Raman scattering,” Opt. Lett. 39(9), 2573–2575 (2014). [CrossRef]  

11. S. M. Pershin, V. N. Lednev, R. N. Yulmetov, V. K. Klinkov, and A. F. Bunkin, “Transparent material thickness measurements by Raman scattering,” Appl. Opt. 54(19), 5943–5948 (2015). [CrossRef]  

12. J. Schmidhuber, “Deep learning in neural networks: An overview,” Neural Networks 61, 85–117 (2015). [CrossRef]  

13. Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature 521(7553), 436–444 (2015). [CrossRef]  

14. F. Jia, Y. Lei, J. Lin, X. Zhou, and N. Lu, “Deep neural networks: A promising tool for fault characteristic mining and intelligent diagnosis of rotating machinery with massive data,” Mech. Syst. Signal. Pr. 72–73, 303–315 (2016). [CrossRef]  

15. J. Liu, M. Osadchy, L. Ashton, M. Foster, C. J. Solomon, and S. J. Gibson, “Deep convolutional neural networks for Raman spectrum recognition: a unified solution,” Analyst 142(21), 4067–4074 (2017). [CrossRef]  

16. S. Krauß, R. Roy, H. Yosef, T. Lechtonen, S. El-Mashtoly, K. Gerwert, and A. Mosig, “Hierarchical deep convolutional neural networks combine spectral and spatial information for highly accurate Raman-microscopy-based cytopathology,” J. Biophotonics 11(10), e201800022 (2018). [CrossRef]  

17. X. Fan, W. Ming, H. Zeng, Z. Zhang, and H. Lu, “Deep learning-based component identification for the Raman spectra of mixtures,” Analyst 144(5), 1789–1798 (2019). [CrossRef]  

18. X. Chen, L. Xie, Y. He, T. Guan, X. Zhou, B. Wang, G. Feng, H. Yu, and Y. Ji, “Fast and accurate decoding of Raman spectra-encoded suspension arrays using deep learning,” Analyst 144(14), 4312–4319 (2019). [CrossRef]  

19. H. M. Fayek, M. Lech, and L. Cavedon, “Evaluating deep learning architectures for Speech Emotion Recognition,” Neural Networks 92, 60–68 (2017). [CrossRef]  

20. Y. Rivenson, Z. Göröcs, H. Günaydin, Y. Zhang, H. Wang, and A. Ozcan, “Deep learning microscopy,” Optica 4(11), 1437–1443 (2017). [CrossRef]  

21. S. Jiang, J. Liao, Z. Bian, K. Guo, Y. Zhang, and G. Zheng, “Transform- and multi-domain deep learning for single-frame rapid autofocusing in whole slide imaging,” Biomed. Opt. Express 9(4), 1601–1612 (2018). [CrossRef]  

22. Z. Ren, Z. Xu, and E. Y. Lam, “Learning-based nonparametric autofocusing for digital holography,” Optica 5(4), 337–344 (2018). [CrossRef]  

23. X. Yuan and Y. Pu, “Parallel lensless compressive imaging via deep convolutional neural networks,” Opt. Express 26(2), 1962–1977 (2018). [CrossRef]  

24. P. Zelger, K. Kaser, B. Rossboth, L. Velas, G. J. Schuetz, and A. Jesacher, “Three-dimensional localization microscopy using deep learning,” Opt. Express 26(25), 33166–33179 (2018). [CrossRef]  

25. G. Zhang, T. Guan, Z. Shen, X. Wang, T. Hu, D. Wang, Y. He, and N. Xie, “Fast phase retrieval in off-axis digital holographic microscopy through deep learning,” Opt. Express 26(15), 19388–19405 (2018). [CrossRef]  

26. S. Purandare, J. L. Zhu, R. J. Zhou, G. Popescu, A. Schwing, and L. L. Goddard, “Optical inspection of nanoscale structures using a novel machine learning based synthetic image generation algorithm,” Opt. Express 27(13), 17743–17762 (2019). [CrossRef]  

27. K. He, X. Zhang, S. Ren, and J. Sun, “Identity Mappings in Deep Residual Networks,” in Computer Vision - Eccv 2016, Pt Iv, B. Leibe, J. Matas, N. Sebe, and M. Welling, eds. (Academic, 2016), pp. 630–645.

28. K. He, X. Zhang, S. Ren, J. Sun, and Ieee, “Deep Residual Learning for Image Recognition,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2016), pp. 770–778.

29. K. Zhang, W. Zuo, Y. Chen, D. Meng, and L. Zhang, “Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising,” in Proceedings of IEEE Transactions on Image Processing (IEEE, 2017), pp. 3142–3155.

30. P. Murugan and S. Durairaj, “Regularization and optimization strategies in deep convolutional neural network,” arXiv preprint arXiv:1712.04711 (2017).

31. N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A Simple Way to Prevent Neural Networks from Overfitting,” J. Mach. Learn. Res. 15, 1929–1958 (2014).

32. S. Panchapagesan, M. Sun, A. Khare, S. M. A. Mandal, B. Hoffineister, S. Vitaladevuni, and A. Int Speech Commun, “Multi-task learning and Weighted Cross-entropy for DNN-based Keyword Spotting,” in 17th Annual Conference of the International Speech Communication Association (2016), pp. 760–764.

33. I. Durickovic, R. Claverie, P. Bourson, M. Marchetti, J. M. Chassot, and M. D. Fontana, “Water-ice phase transition probed by Raman spectroscopy,” J. Raman Spectrosc. 42(6), 1408–1412 (2011). [CrossRef]  

34. B. Lienert, J. Porter, and S. K. Sharma, “Simultaneous measurement of spectra at multiple ranges using a single spectrometer,” Appl. Opt. 48(24), 4762–4766 (2009). [CrossRef]  

35. S. Pershin, A. Lyash, V. Makarov, K. Hamal, I. Prochazka, and B. Sopko, “Multilayer cloud monitoring by micro-Joule lidar based on photon counting receiver and diode laser,” Proc. SPIE 7355, 73550S (2009). [CrossRef]  

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (6)

Fig. 1.
Fig. 1. Experimental setup for ICM: M1 and M2 represent the mirrors; BE the beam expander; BS the beam splitter; OL the objective lens; and CL the collection lens.
Fig. 2.
Fig. 2. Basic architecture of the improved DRN. In each layer, a residual convolutional layer and a max pooling are involved, in which “×3” on the top of “Block” indicates repeating the block three times. “FC1” and “FC2” denote fully connected layers, “Dropout” represents the dropout layer. In the Feature Extraction, the kernel size and depth are demonstrated at the bottom of each layer, in which “depth_n” (n = 1, 2, …, 7) means 32, 32, 64, 64, 128, 128, and 256, respectively. In the Regression, the input Raman signals are predicted with a response, denoting the estimated position of the ice–water interface.
Fig. 3.
Fig. 3. Framework of the max pooling.
Fig. 4.
Fig. 4. Typical Raman spectra of water and floating ice. (a) before ice insertion, (b) after ice insertion, in which “+” indicates the ice spectrum obtained above the ice-water interface, and “−“ indicates the water spectrum obtained beneath the ice-water interface.
Fig. 5.
Fig. 5. Training loss during the training process.
Fig. 6.
Fig. 6. ICM using (a) GFM and (b) the proposed DRN

Tables (1)

Tables Icon

Table 1. Comparison of the measurement error between DRN and GFM

Equations (5)

Equations on this page are rendered with MathJax. Learn more.

c j ( k ) = R [ i = 1 K ( k 1 ) c i ( k 1 ) ω i j ( k ) + b j ( k ) ]
y = F ( x ) + x
L ( y , y r ) = 1 N { i = 1 N y r _ i × log [ softmax ( y i ) ] }
softmax ( y i ) = e y i j = 1 N e y j
L o s s = 1 M [ j = 1 M L ( y j , y r j ) ] + α | | W | | 2 2
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.