Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Fast full-color computational imaging with single-pixel detectors

Open Access Open Access

Abstract

Single-pixel detectors can be used as imaging devices by making use of structured illumination. These systems work by correlating a changing incident light field with signals measured on a photodiode to derive an image of an object. In this work we demonstrate a system that utilizes a digital light projector to illuminate a scene with approximately 1300 different light patterns every second and correlate these with the back scattered light measured by three spectrally-filtered single-pixel photodetectors to produce a full-color high-quality image in a few seconds of data acquisition. We utilize a differential light projection method to self normalize the measured signals, improving the reconstruction quality whilst making the system robust to external sources of noise. This technique can readily be extended for imaging applications at non-visible wavebands.

© 2013 Optical Society of America

1. Introduction

Over the past two decades there has been considerable research in ghost imaging and computational imaging with single-pixel detectors. Although both research areas have different origins, there are instances in which both are closely related. Initially, ghost imaging relied on the use of two correlated light fields and two photodetectors to produce an image; one detector with no spatial resolution used to collect one light field which had previously interacted with an object, and a second detector with high spatial resolution used to collect the other correlated light which never interacts with the object. Neither detector alone is capable of imaging the object, however combining the measurements made by both detectors can be used to produce an image.

Early demonstrations of ghost imaging utilized the entanglement arising from spontaneous parametric down-conversion to produce the correlated light fields, an approach often referred to as quantum ghost imaging [1, 2]. The presence of entanglement led many to interpret this as fundamentally quantum behavior. However, there have since been demonstrations using a pseudothermal light source [37], whereby laser light is propagated through a ground glass diffuser to produce a speckle field, after which a beam splitter makes a correlated copy of the field, a technique commonly termed classical ghost imaging. For both classical and quantum ghost imaging approaches, the field-of-view, spatial resolution, contrast and signal-to-noise ratio of an image can be described by semi-classical photodetection theory [8].

Subsequently, the need for a beam splitter was removed by using a spatial light modulator capable of generating programmable light field to illuminate the object. Controlling both the intensity and phase, allows the intensity structure to be calculated at any plane and stored in computer memory rather than measured on a detector with high spatial resolution. This simplified experimental approach is known as computational ghost imaging [9] and can be performed on systems with and without lenses between the source and object. When an imaging system is employed, any phase information becomes redundant, since only the intensity correlations of the light reflected or transmitted from an object are used to produce an image. Within the signal processing community this latter approach is referred to as a single-pixel camera employing structured illumination [10].

Interestingly, for imaging systems based on structured illumination, the position of the single-pixel detector has been shown to determine the apparent illumination direction for the scene [11]. Moreover, by employing multiple single-pixel devices in different positions, variations in the image shading profiles allow for accurate three-dimensional images to be retrieved [12].

The use of single-pixel detectors as imaging devices has been studied extensively, in particular within the field of compressed sensing, which can be used to exploit the sparseness in natural images to significantly reduce the amount of information that is needed to reproduce it. Indeed, it is this feature of natural images that is at the heart of well-known lossy image compression algorithms, such as JPEG [13].

In this work we present a computational imaging system that can produce full-color 2D images of 3D scenes containing multiple objects within a few seconds. Our system makes use of structured illumination and three single-pixel detectors to simultaneously obtain the red, green and blue color planes. We compare color images that are reconstructed when an iterative algorithm and a compressed sensing technique is applied. The latter approach is shown to yield high-quality full-color images from sub-Nyquist number of measurements.

2. Experimental setup

The setup used in this experiment is shown in Fig. 1. In contrast to classical ghost imaging, this approach relies on using a digital light projector (DLP) to provide spatially incoherent binary structured illumination, which is imaged to the plane of the scene with a 55mm lens. The DLP contains a digital micro-mirror device (DMD) and three colored (red, green and blue) light emitting diodes (LEDs). For each incident pattern the total intensity reflected from the scene is directed onto a composite dichroic beamsplitter (X-Cube) using a large collection lens. The dichroic beam splitter is used to spectrally filter red, green and blue light towards different outputs and allow subsequent measurement on three unfiltered single-pixel photodetectors, as shown in the callout of Fig. 1. The computer records all three photodetector signals via an analogue-to-digital converter, which are then be used for image reconstruction using an appropriate algorithm.

 figure: Fig. 1

Fig. 1 Schematic of the experimental system for imaging a colored 3D scene. A dichroic beamsplitter (X-Cube) spectrally separates white light to three outputs (red, green and blue) where three unfiltered photodiodes measure the reflected intensity for each incident pattern from a digital light projector. The measured signals and projected patterns are used in an computer algorithm to reconstruct an image of the scene.

Download Full Size | PDF

In order to obtain images from the system quickly, it is necessary to rapidly illuminate a scene with different structured light fields. The DMD contained in the light projector has a maximum array switching rate of 1440Hz, determined by the 60 Hz frame rate and 24 bit-planes (used for color depth) of the DLP. As we require only binary illumination patterns, we operate the projector with all three LEDs permanently on and can take advantage of the color planes to display 24 different binary images per frame.

One main challenge when displaying at this rate is overcoming the various latencies involved in the computer graphics pipeline. This is addressed by adding in several synchronization processes during data acquisition. One such process introduces a ‘flash’ bit-plane in the sequence of images which make up one whole frame. This is achieved by turning all mirrors to the on position and then all mirrors to the off position. The remaining 22 bit-planes are then used to display random patterns for the reconstruction process. Additionally, every 50th full frame is used as a synchronization frame, meaning that there are 24 alternating black and white bit-planes. This allows robust correlation of the frames to the measured signals and ensures that only data which is found between two matching synchronisation frames is actually used in the reconstruction algorithm or saved for post processing.

An important part of keeping the system synchronized is ensuring that a new frame is displayed every time the projector refreshes. This is achieved through the use of an OpenGL program [14] that only changes the image when the projector refreshes. The OpenGL program is called from a dedicated loop in the LabVIEW control program, such that it always runs at 60Hz. If pattern generation or acquisition runs at a slower rate than this, patterns are displayed twice, and the acquisition code knows to ignore repeated patterns. Similarly, data acquisition is run from a dedicated loop to allow it to be continuously acquired; separate loops split this data into frame-sized chunks and match it up with the patterns displayed on the DMD. LabVIEW’s queue structures are used to pass data asynchronously between loops, and LabVIEW’s native parallelism allows separate loops to run on different processor cores for increased speed. Within each frame, the 22 bit-planes that are not used for synchronization are split into 11 pairs. Each pair consists of a pattern and its inverse. This allows us to make differential measurements, analogous to lock-in detection at 720Hz. Differential detection significantly reduces the influence of noise such as ambient light fluctuations on the measurement. Overall, the system is able to display and measure approximately 650 unique patterns per second, once differential measurement and synchronisation are taken into account.

3. Image reconstruction

When performing imaging with single-pixel detectors, there are two main types of reconstruction algorithms that can be used for processing the acquired data. Iterative algorithms make a refined estimate of the scene after each new measurement, while inversion algorithms utilize an entire data set in a bulk process to find the best solution for a set of unknowns. For both types of algorithms, the resolution of the final reconstruction is N = x × y, where x and y is the number of pixels in x and y dimensions used for illumination. For each iteration, i, a unique 2D intensity pattern Ii(x, y) is projected onto the object and the corresponding reflected intensity (voltage signal), at spectral frequency μ, is measured for each single-pixel photodetector, Sμi, thus we can write

Sμi=Ii'(x,y)Rμ(x,y)dxdy,
where Rμ(x, y) is the reflection function of the scene for a spectral frequency μ and I′i(x, y) is the 2D pattern after propagation. In our case since the modulator is imaged onto the plane of the object, I′i(x, y) = Ii(x, y). Provided that the entire scene is located within the detectors field of view, the signal measured is directly proportional to the projection of the scene reflectivity onto each illumination pattern, and can effectively be used as a weighting factor in the image reconstruction algorithm.

3.1. Iterative image reconstruction

For iterative image reconstructions it has been shown that normalization of the photodetector signals leads to an improvement in the overall image reconstruction [15, 16]. In this experiment we perform a normalization to signals by maintaining an equal black/white ratio in each illuminating pattern and acquire the differential signal between consecutive positive/negative patterns. The iterative algorithm we employ in our system is known as a traditional ghost imaging algorithm, defined by

Oμ(x,y)=(SμSμ)(I(x,y)I(x,y)),
where Oμ(x, y) is the estimate of the scene and < ... > denotes an ensemble average. In this experiment the average differential signal tends to zero, as does each pixel in the average differential pattern, thus Eq. 2 can be re-written as
Oμ(x,y)=1Mi=1MSμiIi(x,y),
for M samples. The final result from our iterative reconstruction is shown in Fig. 2 for approximately 1 million measurements. The full-color image is obtained by combining the final three images reconstructed from each detector, corresponding to the red, green and blue color channels.

 figure: Fig. 2

Fig. 2 (a) Full-color iterative reconstruction of a 3D scene, obtained by combining the separate reconstructions derived from three photodetectors sensitive to (b) red, (c) green and (d) blue color channels. The illumination patterns and final reconstruction utilized a 256 × 128 pixel resolution and required > 20× Nyquist number of measurements. The 3D scene had dimensions of approximately 20cm × 10cm and was located approximately 1m from the digital light projector.

Download Full Size | PDF

3.2. Inversion image reconstruction

For inversion algorithms we can reshape each 2D pattern, as a 1D array, Ii, and produce a measurement matrix, I, containing all projected patterns, such that

I=[I1I2IM]T.
Similarly, the measured signals can written as a column vector
Sμ=[Sμ1Sμ2SμM]T,
such that the problem can be realised as a set of linear equations given by
IOμ=Sμ,
where Oμ is a set of unknowns, which once recovered can be reshaped to Oμ(x, y), representing the estimate of the scene at spectral frequency μ. In the case where MN, this problem can be solved by least-squares methods. However, as the image resolution increases, the size of I makes performing this reconstruction computationally intensive, while using fewer measurements than the resolution size to address this issue results in an ill-conditioned problem and the image reconstruction quality rapidly decreases.

3.3. Compressive sensing

To reduce the number of measurements required for faithful image reconstruction we can instead employ compressive sensing techniques [17, 18] for which a variety of different approaches exist. In this experiment we make use of the 1 magic toolbox for the Matlab programming language (which is available at www.l1-magic.org). To employ this technique the system must be represented in an appropriate sparse basis, therefore we perform a 1D discrete cosine transform (DCT) on each reshaped pattern Ii such that IIDCT. The ill-conditioned problem can then be expressed as

IDCTO*μ=Sμ,
where O*μ is a 1D vector containing the set of unknowns, for the scene at spectral frequency μ, in DCT space. A solution for O*μ is recovered when minimizing
12SμIDCTO*μ2+λ|O*μ|,
by using preconditioned conjugate gradients methods, where λ is a regularisation parameter [19] and was assigned a value of 0.01 in this experiment for optimum performance. Performing an inverse DCT on O*μ and with appropriate reshaping results in a solution for the scene, Oμ(x, y).

By utilising these compressed sensing techniques the reconstructed image quality can be maintained, even with a highly ill-conditioned system of equations by taking advantage of the sparsity in natural images. In Fig. 3 we compare the iterative algorithm and compressed sensing technique for reconstructing a colored scene with resolution of 128 × 64 pixels, for an increasing number of iterations. When using compressed sensing techniques a high quality image can be observed for 6000 iterations (less than 10 seconds of acquisition), which corresponds to approximately 75% of the Nyquist limit. Compared with the iterative algorithm for the equivalent number of iterations we observe both increased image quality and increased contrast as expected.

 figure: Fig. 3

Fig. 3 Comparison of colored image reconstructions with 128 × 64 pixel resolution for increasing sample size using an iterative and a compressive algorithm. For 8000 (approximately Nyquist) measurements the acquisition time is less than 12 seconds.

Download Full Size | PDF

4. Conclusions

In conclusion we have shown that computational imaging with three single-pixel detectors can be used to produce full-color images of large scenes in just a few seconds of data acquisition. The use of a digital light projector and a suitable computer algorithm allows rapid structured illumination and hence short acquisition times, showing promise for a range of alternative imaging applications. Additionally by employing compressive sensing techniques we have shown significant improvement to the image quality when the number of measurements is below the Nyquist limit.

The low cost of additional photodetectors and the large operating bandwidth afforded by DMD technology opens a range of alternative imaging solutions, such as hyperspectral imaging and particularly for imaging at wavelengths where CCD or CMOS imaging technology is limited. In addition, the use of single-pixel photodetectors may offer low data requirements for sending images, provided the patterns used to generate them are known by both parties. We have demonstrated a pattern projection rate of approximately 650Hz, however in principle DMD technology can display in excess of 20kHz, which could reduce acquisition times further.

Acknowledgments

M. J. P. would like to thank the Royal Society, the Wolfson Foundation and DARPA. We gratefully acknowledge financial support from the EPSRC (Grant EP/I012451/1).

References and links

1. T. B. Pittman, Y. H. Shih, D. V. Strekalov, and A. V. Sergienko, “Optical imaging by means of two-photon quantum entanglement,” Phys. Rev. A 52, R3429–R3432 (1995). [CrossRef]   [PubMed]  

2. D. V. Strekalov, A. V. Sergienko, D. N. Klyshko, and Y. H. Shih, “Observation of two-photon “ghost” interference and diffraction,” Phys. Rev. Lett. 74, 3600–3603 (1995). [CrossRef]   [PubMed]  

3. R. S. Bennink, S. J. Bentley, and R. W. Boyd, ““Two-photon” coincidence imaging with a classical source,” Phys. Rev. Lett. 89, 113601 (2002). [CrossRef]  

4. A. Gatti, E. Brambilla, M. Bache, and L. A. Lugiato, “Ghost imaging with thermal light: Comparing entanglement and classical correlation,” Phys. Rev. Lett. 93, 093602 (2004). [CrossRef]   [PubMed]  

5. A. Gatti, E. Brambilla, M. Bache, and L. A. Lugiato, “Correlated imaging, quantum and classical,” Phys. Rev. A 70, 013802 (2004). [CrossRef]  

6. A. Valencia, G. Scarcelli, M. D’Angelo, and Y. Shih, “Two-photon imaging with thermal light,” Phys. Rev. Lett. 94, 063601 (2005). [CrossRef]   [PubMed]  

7. F. Ferri, D. Magatti, A. Gatti, M. Bache, E. Brambilla, and L. A. Lugiato, “High-resolution ghost image and ghost diffraction experiments with thermal light,” Phys. Rev. Lett. 94, 183602 (2005).

8. J. H. Shapiro and R. W. Boyd, “The physics of ghost imaging,” Quantum Inf. Process. 11, 949–993 (2012). [CrossRef]  

9. J. H. Shapiro, “Computational ghost imaging,” Phys. Rev. A 78, 061802 (2008). [CrossRef]  

10. M. Duarte, M. Davenport, D. Takhar, J. Laska, T. Sun, K. Kelly, and R. Baraniuk, “Single-pixel imaging via compressive sampling,” IEEE Signal Proc. Mag. 25, 83–91 (2008). [CrossRef]  

11. P. Sen, B. Chen, G. Garg, S. R. Marschner, M. Horowitz, M. Levoy, and H. P. A. Lensch, “Dual photography,” ACM Trans. Graph. 24, 745–755 (2005). [CrossRef]  

12. B. Sun, M. P. Edgar, R. Bowman, L. E. Vittert, S. Welsh, A. Bowman, and M. J. Padgett, “3D computational imaging with single-pixel detectors,” Science 340, 844–847 (2013). [CrossRef]   [PubMed]  

13. G. K. Wallace, “The jpeg still picture compression standard,” Commun. ACM 34, 30–44 (1991). [CrossRef]  

14. D. Preece, R. Bowman, A. Linnenberger, G. Gibson, S. Serati, and M. Padgett, “Increasing trap stiffness with position clamping in holographic optical tweezers,” Opt. Express 17, 22718–22725 (2009). [CrossRef]  

15. F. Ferri, D. Magatti, L. A. Lugiato, and A. Gatti, “Differential ghost imaging,” Phys. Rev. Lett. 104, 253603 (2010). [CrossRef]   [PubMed]  

16. B. Sun, S. S. Welsh, M. P. Edgar, J. H. Shapiro, and M. J. Padgett, “Normalized ghost imaging,” Opt. Express 20, 16892–16901 (2012). [CrossRef]  

17. O. Katz, Y. Bromberg, and Y. Silberberg, “Compressive ghost imaging,” Appl. Phys. Lett 95, 131110 (2009). [CrossRef]  

18. D. Donoho, “Compressed sensing,” IEEE Trans. Inf. Theory 52, 1289–1306 (2006). [CrossRef]  

19. K. Koh, S.-J. Kim, and S. P. Boyd, “An interior-point method for large-scale l1-regularized logistic regression.” J. Mach. Learn. Res. 8, 1519–1555 (2007).

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (3)

Fig. 1
Fig. 1 Schematic of the experimental system for imaging a colored 3D scene. A dichroic beamsplitter (X-Cube) spectrally separates white light to three outputs (red, green and blue) where three unfiltered photodiodes measure the reflected intensity for each incident pattern from a digital light projector. The measured signals and projected patterns are used in an computer algorithm to reconstruct an image of the scene.
Fig. 2
Fig. 2 (a) Full-color iterative reconstruction of a 3D scene, obtained by combining the separate reconstructions derived from three photodetectors sensitive to (b) red, (c) green and (d) blue color channels. The illumination patterns and final reconstruction utilized a 256 × 128 pixel resolution and required > 20× Nyquist number of measurements. The 3D scene had dimensions of approximately 20cm × 10cm and was located approximately 1m from the digital light projector.
Fig. 3
Fig. 3 Comparison of colored image reconstructions with 128 × 64 pixel resolution for increasing sample size using an iterative and a compressive algorithm. For 8000 (approximately Nyquist) measurements the acquisition time is less than 12 seconds.

Equations (8)

Equations on this page are rendered with MathJax. Learn more.

S μ i = I i ' ( x , y ) R μ ( x , y ) d x d y ,
O μ ( x , y ) = ( S μ S μ ) ( I ( x , y ) I ( x , y ) ) ,
O μ ( x , y ) = 1 M i = 1 M S μ i I i ( x , y ) ,
I = [ I 1 I 2 I M ] T .
S μ = [ S μ 1 S μ 2 S μ M ] T ,
IO μ = S μ ,
I D C T O * μ = S μ ,
1 2 S μ I D C T O * μ 2 + λ | O * μ | ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.