Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Video stabilization in atmosphere turbulent conditions based on the Laplacian-Riesz pyramid

Open Access Open Access

Abstract

Video stabilization in atmosphere turbulent conditions is aimed at removing spatiotemporally varying distortions from video recordings. Conventional shaky video stabilization approaches do not perform effectively under turbulent circumstances due to the erratic motion common to those conditions. Using complex-valued image pyramids, we propose a method to mitigate this erratic motion in videos. First, each frame of a video is decomposed into different spatial frequencies using the Laplacian pyramid. Second, a Riesz transform is adopted to extract the local amplitude and the local phase of each sub-band. Next, low-pass filters are designed to attenuate the local amplitude and phase variations to remove turbulence-induced distortions. Experimental results show that the proposed approach is efficient and provides stabilizing video in atmosphere turbulent conditions.

© 2016 Optical Society of America

1. Introduction

The non-uniform distribution of temperature yields random atmosphere refractive-index fluctuations, resulting in optical turbulence that distorts the optical wave front. This phenomenon causes spatiotemporal distortions and spatial image blur in long-distance imaging. Techniques to mitigate optical turbulence have resulted in image degradation, such as the hardware-based adaptive optics approach and the software-based numerical processing approach. The adaptive optics approach [1] is applied in large telescopes and reduces the turbulence effects in real time. However, its working mode makes this approach unsuitable for horizontal imaging in atmospheric boundary layers.

The numerical processing approach can be divided into two classes. The first is video-to-image type methods. These methods reconstruct a single high-quality image from atmospherically distorted video, such as lucky imaging [2], speckle imaging [3, 4], image restoration methods [5, 6], non-rigid registration methods [7], registration and fusion hybrid methods [8]. The other class consists of video-to-video type methods, namely video stabilization methods, such as the super-resolution method [9], the low-rank matrix decomposition based method [10], the Sobolev sharpening flow based method [11] and the phase based method [12]. In this paper, we focus on investigating video stabilization problems. The super-resolution method first generates a reference frame using a temporal pixel-wise median over a block of frames. Video stabilization is obtained by computing a map of pixel displacements between every frame and its reference frame. To some extent, the effects of stabilization depend on elastic local registration results. The low-rank matrix decomposition method decomposes the turbulence sequence into three components: the background, the turbulence and the object. However, matrix decomposition is time consuming and complicated for high-resolution imaging. The Sobolev sharpening flow-based method proposes a PDE model for video stabilization, and numerical acceleration techniques are used to improve efficiency. However, the target properties in infrared images and the contrasts in panchromatic images are affected using the Sobolev sharpening flow, and details cannot be restored in the event of destructive turbulent degradation. The phase-based method alleviates turbulence distortion through local phase temporal filtering within a complex-valued steerable pyramid. However, the results with this method can be affected by noisy disturbances because the method did not process the amplitude variations.

Wave-front angle-of-arrival (AOA) fluctuations are the phase related parameters of optical wave propagating through atmospheric turbulence, which have two primary degrading effects on recorded video sequences: spatiotemporal distortions and long-exposure imaging blur. The phase of a signal implies the structure and motion information [13]; therefore, phase information has been widely applied in signal processing and image processing fields [8, 12, 14–17]. In one-dimensional (1-D) signal processing, the Hilbert transform based analytic signal is used as a complex-valued representation of a 1-D signal for local frequency analysis. For 2-D signal processing problems, such as image processing, the Riesz transform based monogenic signal [18] has been proposed for analyzing the local amplitude and the local phase.

In this paper, we present an image complex pyramid representation, the Laplacian-Riesz pyramid, which is used for video stabilization in turbulence. The Laplacian pyramid is characterized by its compact structure and perfect reconstruction [19], and it is completely built in the spatial domain; consequently, it possesses high computational efficiency. The Riesz transform is a steerable Hilbert transformer and can compute phase-shift and translate image features in the direction of the dominant orientation at every pixel [17]. Therefore, in this work we combine the Laplacian pyramid with the Riesz transform to form a Laplacian-Riesz pyramid (LRP) for image representation. Based on this complex pyramid representation, local phase and amplitude variations of the coefficients of a Laplacian-Riesz pyramid over time can be extracted. Local phase variations correspond to turbulent-induced wiggles. Additionally, AOA effects also result in local amplitude fluctuations. Therefore, low-pass filters are used to temporally attenuate these local phase and amplitude variations to remove distracting changes. The high spatial frequencies are undesirable for the temporal filtering due to camera sensor and quantization noise. Therefore, we decompose each frame into the pyramid structure first for the purpose of temporal processing.

The paper is organized as follows: Section 2 describes the Laplacian-Riesz pyramid-based video stabilization framework in detail. Experimental results are given in Section 3 to show the performance compared with other methods; finally, we conclude and discuss the direction of future research in Section 4.

2. Laplacian-Riesz pyramid based video stabilization method

The proposed video stabilization framework contains three steps (see the diagram in Fig. 1), (1) the Laplacian-Riesz pyramid representation of each frame of a video; (2) the temporal filtering of local amplitude and local phase; (3) the reconstruction of the corrected frame.

 figure: Fig. 1

Fig. 1 Block diagram for the proposed video stabilization framework.

Download Full Size | PDF

2.1 Laplacian-Riesz pyramid

2.1.1 Laplacian pyramid

A Laplacian pyramid [19] is the subtraction between the lower level’s up-sampling and the higher level in a Gaussian pyramid; this means the Laplacian pyramid decreases the correlation between each level. The Laplacian pyramid forms a series of band-pass components and a low-pass component of the original image. Without losing generality, we used a 5 × 5 Gaussian kernel w(.) as the low-pass filter to build the pyramid. We set the standard deviation of w(.) to 2 because it provides adequate filtering with low reconstruction errors. When the turbulence effects are particularly severe, the standard deviation should be larger to avoid grid noise.

The original frame I convolves with w to get the first level g1 of the Gaussian pyramid, which is treated as the highest level here. The function REDUCE denotes a down-sample and low-pass process, which turns a higher sub-band gk1 with a size of 2Ck×2Rk into a lower sub-band gk with a size of Ck×Rk

gk(i,j)=REDUCE(gk1)=m=22n=22w(m,n)gk1(2i+m,2j+n)
where 1iCk, 1jRk.

Inversely, the function EXPAND denotes the up-sample and interpolation operation, which expands the lower sub-band gk with a size of Ck×Rk to a higher sub-band gk1 with a size of 2Ck×2Rk by interpolating new values between the given values:

gk,2(i,j)=EXPAND(gk)=4m=22n=22w(m,n)gk(im2,jn2)
where 1i2Ck, 1j2Rk. Note that the pixel coordinates must be integers. Only when (im)/2 and (jn)/2 are integers, will the term gk be included in this sum. Equation (2) gives the value to non-included pixels by convolving the 5 × 5 neighborhood values with the weight matrix 4w.

Suppose each sub-band of the Laplacian pyramid is L{k}, k=1,2,,N. N indicates the maximum level. The Laplacian pyramid can then be built as follows:

L{1}=IEXPAND(g2)=Ig2,2
L{k}=gkEXPAND(gk+1)=gkgk+1,2for2k<N
L{N}=gN

Equations (3)-(5) show that the Laplacian pyramid can be reconstructed exactly to recover the original image. gN1 can be computed by adding L{N1} to the EXPAND of L{N}. Then, gN2 is yielded by adding the EXPAND of gN1 to L{N2}. The rest can be done in the same manner:

gk1=L{k1}+EXPAND(gk)

Finally, a reconstructed image Iout is recovered from the pyramid:

Iout=L{1}+EXPAND(g2)

2.1.2 Riesz transform

The Riesz transform is the multi-dimensional extension of the Hilbert transform. The Riesz operator’s frequency response is jω/ω. That is, there are N Riesz filters for an N-dimension signal, and when N = 1 the Riesz transform is equal to the Hilbert transform. As for a 2D signal I(x), x=[x,y], the Riesz transform is expressed as:

(x)=(R1(x)R2(x))=(hxI(x)hyI(x))
where the filter hx=x/(2πx3) and hy=y/(2πx3) correspond to the 2D frequency responses Hx(ω)=jωx/ω and Hy(ω)=jωy/ω, ω=[ωx,ωy], respectively.

The frequency response of the Riesz operator indicates that the Riesz transform is a 90-degree phase-shift system. The phase-shift property enables the Riesz transform to compute the local amplitude and local phase along the dominant orientation θ at every point of an image I.

The input I and its Riesz transform (R1,R2) together form a triple that can output the local amplitude A and local phase φ:

I=Acos(φ),R1=Asin(φ)cos(θ),R2=Asin(φ)sin(θ)

The local amplitude A implies the local intensity and energy, and it is defined as:

A=I2+R12+R22

The local phase φ contains the structural and motion information, φ is defined as:

φ=cos1(I/A)

2.1.3 Laplacian-Riesz pyramid

After building the Laplacian pyramid, we performed the Riesz transform to each sub-band L{k}. The LRP can then be created: {R1{k},R2{k}}. There are two components on each level of the LRP. Figures 2(a) and 2(b) illustrate the LRP structure. With Eqs. (10) and (11), the LRP provides a multi-resolution method to analyze the local amplitude A{k} and local phase φ{k}, as is shown in Fig. 2(c).

 figure: Fig. 2

Fig. 2 LRP and Local information (a) Input image, (b) LRP, (c) Local amplitude and local phase.

Download Full Size | PDF

2.2 Temporal filtering of the local amplitude and local phase

The impact of turbulence on imaging appears as pixel intensity blinking and pixel wiggles, which can be represented and analyzed by the local amplitude and the local phase of the Riesz transform. In this paper, the Riesz transform is applied to reveal the turbulence distortion effects.

Assume we are given an image sequence without turbulence effects: I(x,y,t), where t denotes the sequence frame number. In Eq. (9), I(x,y,t)=A(x,y,t)cosφ(x,y,t), A(x,y,t) and φ(x,y,t) are the amplitude and the phase of the Riesz transform of non-turbulence videos. A(x,y,t) and φ(x,y,t) are the temporal constants in areas that are not influenced by turbulence. Therefore, we use A(x,y) and φ(x,y) to replace A(x,y,t) and φ(x,y,t).

When the image sequence is influenced by turbulence, the degraded effects cause the local brightness variation, which is represented by Δ(x,y,t), and also create a local ‘dance’ δ(x,y,t). Turbulence effects can be expressed as:

ITurbu(x,y,t)=[A(x,y)+Δ(x,y,t)]cos[φ(x,y)+δ(x,y,t)]
where ITurbu(x,y,t) indicates each distorted frame. Δ(x,y,t) and δ(x,y,t) represent the turbulence effects on the amplitude and phase from the perspective of temporal varying. Equation. (12) is derived from the turbulence-degraded effects and the Riesz transform. It is also the mathematical foundation of our algorithm.

Using the Riesz transform and Eqs. (10) and (11), the local amplitude A(x,y)+Δ(x,y,t) and local phase φ(x,y)+δ(x,y,t) can be extracted. Figure 3 illustrates a video’s local amplitude and the local phase temporal variance. We found that the local amplitude and local phase in turbulence videos are displayed like high-frequency fluctuations. In this case, low-pass filtering may be reasonable to eliminate the distortions and retain the DC components: A(x,y) and φ(x,y).

 figure: Fig. 3

Fig. 3 The high-frequency fluctuations of turbulence distortion. (a) One frame of Sequence 2. (b) Local amplitude and local phase temporal varying of the green point .

Download Full Size | PDF

In our experiments, the first order Butterworth filter was chosen as the low-pass filter for comparison with the results of [12]. Of course, other low-pass filters are also feasible. In general, the higher the order of a filter, the more accurate the approximation, but the calculation cost also rises. Therefore, a balance must be found between the performance and the cost. For this we chose the first order Butterworth filter.

The first order Butterworth low-pass filter’s frequency response can be expressed as:

HButter(ω)2=11+(ω/ωc)2
where ωc is the cutoff frequency, which is decided by the turbulence strength and stability requirement. Through experimentation we found that the turbulence fluctuation frequency is normally higher than 1Hz. Therefore, ωc can be set to 0.1~0.8Hz for most of the scenes; the stronger the turbulence, the smaller the ωc. The same ωc can be used for the local amplitude and the local phase filters.

The low-pass filtering is carried out by IIR filter, whose difference equation is:

k=0Na[k]out[nk]=k=0Mb[k]input[nk]
where two 2D parameters b and a are generated by the first order Butterworth filter. input[.] denotes the input signal, out[.] denotes the output signal, for the case of first order: M=N=1.

After temporal filtering, the original {A{k},φ{k}} of the distorted frame I is modified to be {yA{k},yφ{k}}. Then the filtered Laplacian pyramid can be expressed as Lout{k}=yA{k}cosyφ{k}. The undistorted frame Iout is rebuilt using Eqs. (6) and (7).

3. Experiments

In this section, we conducted experiments to illustrate the performance of the proposed approach using both simulated and real videos. Results of this video-to-video type approach in [11] and [12] are also shown for comparison (the outputs of method [11] were produced through our own implementation; the outputs of method [12] were generated using the original code [20]). All experiments were performed on a PC with an Intel i5-4460 processor and 8G RAM (the algorithm uses only a single CPU core). We implemented these methods in Matlab 2013b.

3.1 Simulated experiments

To quantitatively evaluate the method performance, we generated a set of turbulence degraded image sequences in horizontal views. This simulation method is a hybrid one [21, 22] that combines the physical models of optical waves under Kolmogorov turbulence with imaging processing algorithms (image convolution and interpolation methods). Simulation uses a latent sharp image (280 × 280) as an input image (as shown in Fig. 4) and produces image sequences with 100 frames degraded by specified turbulence, which includes imaging distance, optics diameter, wavelength, turbulence spectral power law, and turbulence strength. Simulation parameters were designed as: the imaging distance T = 1000m, wavelength λ = 0.5μm, turbulent strength (strong turbulence) Cn2 = 2.5 × 10−13m-2/3, the aperture diameter D = 0.3m. In the following, we used other parameter settings as described in [21]. We refer interested readers to [21, 22] for details.

 figure: Fig. 4

Fig. 4 Latent sharp image used for simulation.

Download Full Size | PDF

The cutoff frequency ωc is set to several values in order to show the effects of this parameter. Figure 5 shows the performance of the different ωc. The yellow line profile in each image is plotted in time on the right (for a fixed x, plot of y vs t). As we can see, the smaller ωc is, the more stable the corresponding result.

 figure: Fig. 5

Fig. 5 The effects of ωc (see Visualization 1) (a) one frame from the simulated video, (b) ωc = 0.1, (c) ωc = 0.3, (d) ωc = 0.5, (e) ωc = 0.8.

Download Full Size | PDF

Figure 6 shows a comparison between the other two methods, and Figs. 6(a) and 6(b) show the outputs of [12] and [11], respectively. As shown [12], effectively alleviates the distortions, but the local brightness is obviously still blinking. The method in [11] can reduce the distortion to some extent, but the result is not stabilized enough or produces cross-color during restoration. Our proposed method outperforms the others.

 figure: Fig. 6

Fig. 6 Simulated comparison results (see Visualization 2) (a) [12] output, (b) [11] output, (c) the proposed output (with ωc = 0.1).

Download Full Size | PDF

3.2 Quantitative comparisons

To quantitatively evaluate the performance, we measured the Peak Signal-to-Noise Ratio (PSNR) between each frame and reference frame. For the simulated video, we used the latent image as the reference; the results are listed in Table 1. For the real videos, we used the average frame of the sequence as the reference. Table 2 reports the comparison results. It is clear that our method obtained state-of-art results and outperformed the other two methods.

Tables Icon

Table 1. Average PSNR (dB) for differentωc values

Tables Icon

Table 2. Comparison of the Average PSNR (dB)

To evaluate the robustness to noise, we added white Gaussian noise with different SNRs to the simulated video. Table 3 shows the performance. Figure 7 shows the visual results.

Tables Icon

Table 3. Performance with white Gaussian noise

 figure: Fig. 7

Fig. 7 Results of additional noises (see Visualization 3, Visualization 4, Visualization 5) (a) one frame of simulated video with noise, (b) one frame of [11] output, (c) one frame of [12] output, (d) one frame of the proposed output.

Download Full Size | PDF

In Table 4, the total consuming times are presented. As shown, the proposed approach is the fastest one.

Tables Icon

Table 4. Comparison of running time (s)

3.3 Real video experiments

This work tested three real sequences distorted by atmospheric turbulence to illustrate the performance of the proposed approach. Sequence 1 consists of 101 frames degraded by strong turbulence effects. Sequences 2 and 3 are two infrared sequences [23]. Sequence 3 contains a moving man and two moving dogs. Figures 8, 9, and 10 show three turbulence mitigation experimental results. The cutoff frequency is set to 0.3Hz for Sequence 1 and 2, 0.8Hz for Sequence 3. Visual effect comparisons show that the proposed approach obtained the best turbulence mitigation result.

 figure: Fig. 8

Fig. 8 Sequence1 stabilization results (see Visualization 6) (a) one frame from sequence 1, (b) [12] output, (c) [11] output, (d) the proposed output.

Download Full Size | PDF

 figure: Fig. 9

Fig. 9 Sequence 2 stabilization results (see Visualization 7) (a) one frame from sequence 2, (b) [12] output, (c) [11] output, (d) the proposed output.

Download Full Size | PDF

 figure: Fig. 10

Fig. 10 Sequence 3 stabilization results (see Visualization 8) (a) one frame from sequence 3, (b) [12] output, (c) [11] output, (d) the proposed output.

Download Full Size | PDF

4. Discussion and conclusions

In this work we proposed a fast method to stabilize video distorted by atmospheric turbulence. In this approach, each frame was decomposed into non-oriental sub-bands with a Laplacian pyramid that was built with spatial Gaussian kernel convolution and difference, then local amplitude and phase along the dominant orientation was computed directly with the Riesz transform. Therefore, the efficiency of the proposed approach is much higher than the phase-based method [12], which used eight orientation complex steerable pyramid representations. Additionally, our method used temporal filtering of the local amplitude and local phase to mitigate the temporal oscillation of atmospheric turbulence. As a result, our method obtained excellent visual quality.

We noted that our method retained the moving objects while stabilizing turbulence distortions in Sequence 3. However, the movement was weakened to some extent because low-pass filtering is harmful to movement. Eliminating the turbulence effects without harming moving objects remains a challenge. Maybe in the future, when turbulence fluctuation frequencies are known, the low-pass filter can be replaced by a band-stop filter to hold back the turbulence effects. Consequently, moving objects would be better preserved.

Temporal oscillation and blurry images are two of the main degradation effects of atmospheric turbulence. Results of video stabilization using the proposed method still suffer from turbulence-induced blur. To obtain a latent sharp frame, we need to know the atmospheric modulation transfer function (MTF) [24]. And then, deconvolution techniques should be used to restore each sharp frame [25, 26]. These remain open problems with plenty of room for further investigation.

References and links

1. R. K. Tyson, Principles of Adaptive Optics (Academic, 1991).

2. Institute of Astronomy, University of Cambridge, “Lucky Imaging”, http://www.ast.cam.ac.uk/~optics/Lucky_Web_Site/index.htm

3. M. Loktev, O. Soloviev, S. Savenko, and G. Vdovin, “Speckle imaging through turbulent atmosphere based on adaptable pupil segmentation,” Opt. Lett. 36(14), 2656–2658 (2011). [CrossRef]   [PubMed]  

4. D. A. Hope, S. M. Jefferies, M. Hart, and J. G. Nagy, “High-resolution speckle imaging through strong atmospheric turbulence,” Opt. Express 24(11), 12116–12129 (2016). [CrossRef]   [PubMed]  

5. D. Li, R. M. Mersereau, and S. Simske, “Atmospheric turbulence-degraded image restoration using principal components analysis,” IEEE Geosci. Remote Sens. Lett. 4(3), 340–344 (2007). [CrossRef]  

6. X. Zhu and P. Milanfar, “Image reconstruction from videos distorted by atmospheric turbulence,” Proc. SPIE 7543(1), 423–426 (2010).

7. R. Gal, N. Kiryati, and N. Sochen, “Progress in the restoration of image sequences degraded by atmospheric turbulence,” Pattern Recognit. Lett. 48, 8–14 (2014). [CrossRef]  

8. N. Anantrasirichai, A. Achim, N. G. Kingsbury, and D. R. Bull, “Atmospheric turbulence mitigation using complex wavelet-based fusion,” IEEE Trans. Image Process. 22(6), 2398–2408 (2013). [CrossRef]   [PubMed]  

9. L. Yaroslavsky, B. Fishbain, G. Shabat, and I. Ideses, “Superresolution in turbulent videos: making profit from damage,” Opt. Lett. 32(20), 3038–3040 (2007). [CrossRef]   [PubMed]  

10. O. Oreifej, X. Li, and M. Shah, “Simultaneous video stabilization and moving object detection in turbulence,” IEEE Trans. Pattern Anal. Mach. Intell. 35(2), 450–462 (2013). [CrossRef]   [PubMed]  

11. Y. Lou, S. H. Kang, S. Soatto, and A. L. Bertozzi, “Video stabilization of atmospheric turbulence distortion,” Inverse Probl. Imaging (Springfield) 7(3), 839–861 (2013). [CrossRef]  

12. N. Wadhwa, M. Rubinstein, F. Durand, and W. T. Freeman, “Phase-based video motion processing,” ACM TOG 32(4), 80 (2013). [CrossRef]  

13. A. V. Oppenheim and J. S. Lim, “The importance of phase in signals,” InProceedings of the IEEE (IEEE, 1981), 69(5), pp, 529–541.

14. M. C. Morrone and R. A. Owens, “Feature detection from local energy,” Pattern Recognit. Lett. 6(5), 303–313 (1987). [CrossRef]  

15. P. Kovesi, “Image features from phase congruency,” J. Computer Vis. Res. 1(3), 1–26 (1999).

16. F. J. Madrid-Cuevas, R. Medina-Carnicer, Á. Carmona-Poyato, and N. L. Fernández-García, “Dominant Points Detection Using Phase Congruence,” Iberian Conference on Pattern Recognition and Image Analysis (Springer Berlin Heidelberg, 2007), pp.138–145. [CrossRef]  

17. N. Wadhwa, M. Rubinstein, F. Durand, and W. T. Freeman, “Riesz pyramids for fast phase-based video magnification,” In proceedings of ICCP, (IEEE, 2014), pp.1–10.

18. M. Felsberg and G. Sommer, “The monogenic signal,” IEEE Trans. Signal Process. 49(12), 3136–3144 (2001). [CrossRef]  

19. P. J. Burt and E. H. Shashua, “The laplacian pyramid as a compact image code,” IEEE Trans. Commun. 31(4), 532–540 (1983). [CrossRef]  

20. N. Wadhwa, M. Rubinstein, F. Durand, and W. T. Freeman, “SIGGRAPH 2013 Phase-Based Video Motion Processing”, http://people.csail.mit.edu/nwadhwa/phase-video/

21. E. Repasi and R. Weiss, “Analysis of Image Distortions by Atmospheric Turbulence and Computer Simulation of Turbulence Effects,” Proc. SPIE 6941, 69410S (2008). [CrossRef]  

22. E. Repasi and R. Weiss, “Computer Simulation of Image Degradations by Atmospheric Turbulence for Horizontal Views,” Proc. SPIE 8014, 80140U (2011). [CrossRef]  

23. O. Haik and Y. Yitzhaky, http://www.ee.bgu.ac.il/~itzik/VideosOE06/

24. I. Dror and N. S. Kopeika, “Experimental comparison of turbulence MTF and aerosol MTF through open atmosphere,” J. Opt. Soc. Am. A 12(5), 970–980 (1995). [CrossRef]  

25. X. Zhu and P. Milanfar, “Removing atmospheric Turbulence via Space-Invariant Deconvolution,” IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 157–170 (2013). [CrossRef]   [PubMed]  

26. P. Ruiz, X. Zhou, J. Mateos, R. Molina, and A. K. Katsaggelos, “Variational Bayesian Blind Image Deconvolution: A review,” Digit. Signal Process. 47, 116–127 (2015). [CrossRef]  

Supplementary Material (8)

NameDescription
Visualization 1: MP4 (1663 KB)      The effects of omega sub c
Visualization 2: MP4 (3175 KB)      Simulated experiments
Visualization 3: MP4 (7395 KB)      SNR 5db
Visualization 4: MP4 (6916 KB)      SNR 15db
Visualization 5: MP4 (6352 KB)      SNR 25db
Visualization 6: MP4 (4747 KB)      Sequence1 stabilization results
Visualization 7: MP4 (4527 KB)      Sequence 2 stabilization results
Visualization 8: MP4 (3107 KB)      Sequence 3 stabilization results

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (10)

Fig. 1
Fig. 1 Block diagram for the proposed video stabilization framework.
Fig. 2
Fig. 2 LRP and Local information (a) Input image, (b) LRP, (c) Local amplitude and local phase.
Fig. 3
Fig. 3 The high-frequency fluctuations of turbulence distortion. (a) One frame of Sequence 2. (b) Local amplitude and local phase temporal varying of the green point .
Fig. 4
Fig. 4 Latent sharp image used for simulation.
Fig. 5
Fig. 5 The effects of ωc (see Visualization 1) (a) one frame from the simulated video, (b) ωc = 0.1, (c) ωc = 0.3, (d) ωc = 0.5, (e) ωc = 0.8.
Fig. 6
Fig. 6 Simulated comparison results (see Visualization 2) (a) [12] output, (b) [11] output, (c) the proposed output (with ωc = 0.1).
Fig. 7
Fig. 7 Results of additional noises (see Visualization 3, Visualization 4, Visualization 5) (a) one frame of simulated video with noise, (b) one frame of [11] output, (c) one frame of [12] output, (d) one frame of the proposed output.
Fig. 8
Fig. 8 Sequence1 stabilization results (see Visualization 6) (a) one frame from sequence 1, (b) [12] output, (c) [11] output, (d) the proposed output.
Fig. 9
Fig. 9 Sequence 2 stabilization results (see Visualization 7) (a) one frame from sequence 2, (b) [12] output, (c) [11] output, (d) the proposed output.
Fig. 10
Fig. 10 Sequence 3 stabilization results (see Visualization 8) (a) one frame from sequence 3, (b) [12] output, (c) [11] output, (d) the proposed output.

Tables (4)

Tables Icon

Table 1 Average PSNR (dB) for differentωc values

Tables Icon

Table 2 Comparison of the Average PSNR (dB)

Tables Icon

Table 3 Performance with white Gaussian noise

Tables Icon

Table 4 Comparison of running time (s)

Equations (14)

Equations on this page are rendered with MathJax. Learn more.

g k ( i,j )=REDUCE( g k1 )= m=2 2 n=2 2 w( m,n ) g k1 ( 2i+m,2j+n )
g k,2 ( i,j )=EXPAND( g k )=4 m=2 2 n=2 2 w( m,n ) g k ( im 2 , jn 2 )
L { 1 } =IEXPAND( g 2 )=I g 2,2
L { k } = g k EXPAND( g k+1 )= g k g k+1,2 for 2k<N
L { N } = g N
g k1 = L { k1 } +EXPAND( g k )
I out = L { 1 } +EXPAND( g 2 )
( x )=( R 1 ( x ) R 2 ( x ) )=( h x I( x ) h y I( x ) )
I=Acos( φ ), R 1 =Asin( φ )cos( θ ), R 2 =Asin( φ )sin( θ )
A= I 2 + R 1 2 + R 2 2
φ= cos 1 ( I/A )
I Turbu ( x,y,t )=[ A( x,y )+Δ( x,y,t ) ]cos[ φ( x,y )+δ( x,y,t ) ]
H Butter ( ω ) 2 = 1 1+ ( ω/ ω c ) 2
k=0 N a[ k ]out[ nk ]= k=0 M b[ k ] input[ nk ]
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.