ARTICLE
Received 12 Sep 2011 | Accepted 13 Feb 2012 | Published 20 Mar 2012 DOI: 10.1038/ncomms1747
The recovery of objects obscured by scattering is an important goal in imaging and has been approached by exploiting, for example, coherence properties, ballistic photons or penetrating wavelengths. Common methods use scattered light transmitted through an occluding material, although these fail if the occluder is opaque. Light is scattered not only by transmission through objects, but also by multiple reection from diffuse surfaces in a scene. This reected light contains information about the scene that becomes mixed by the diffuse reections before reaching the image sensor. This mixing is difcult to decode using traditional cameras. Here we report the combination of a time-of-ight technique and computational reconstruction algorithms to untangle image information mixed by diffuse reection. We demonstrate a three-dimensional range camera able to look around a corner using diffusely reected light that achieves sub-millimetre depth precision and centimetre lateral precision over 40 cm 40 cm40 cm of hidden space.
Recovering three-dimensional shape around a corner using ultrafast time-of-ight imaging
Andreas Velten1, Thomas Willwacher2, Otkrist Gupta1, Ashok Veeraraghavan3, Moungi G. Bawendi4 & Ramesh Raskar1
1 MIT Media Lab, Massachusetts Institute of Technology, 75 Amherst Street, E14-474, Cambridge, Massachusetts 02139, USA. 2 Department of Mathematics, Harvard University, One Oxford Street, Cambridge, Massachusetts 02138, USA. 3 Department of Electrical and Computer Engineering, Rice University, 6100 Main Street, Houston, Texas 77005, USA. 4 Department of Chemistry, Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, Massachusetts 02139, USA. Correspondence and requests for materials should be addressed to R.R. (email: [email protected]).
NATURE COMMUNICATIONS | 3:745 | DOI: 10.1038/ncomms1747 | www.nature.com/naturecommunications
2012 Macmillan Publishers Limited. All rights reserved.
ARTICLE
NATURE COMMUNICATIONS | DOI: 10.1038/ncomms1747
The light detected on an image sensor is composed of direct light, that travels directly from the light source to an object in the line of sight of the sensor, and indirect light that interacts
with other parts of the scene before striking an object in the line of sight. Light from objects outside the line of sight reaches the sensor as indirect light, via multiple reections (or bounces). In conventional imaging, it is difficult to exploit this non-line-of-sight light, if the reections or bounces are diuse.
Line-of-sight time-of-ight information is commonly used in LIDAR (light detection and ranging)1 and two dimensional gated viewing2 to determine the object distance, or to reject unwanted scattered light. By considering only the early ballistic photons from a sample, these methods can image through turbid media or fog3. Other methods, like coherent LIDAR4, exploit the coherence of light to determine the time of ight. However, light that has undergone multiple diuse reections has diminished coherence.
Recent methods in computer vision and inverse light transport study multiple diuse reections in free space. Dual photography5 shows one can exploit scattered light to recover two-dimensional (2D) images of objects illuminated by a structured dynamic light source and hidden from the camera. Time-gated viewing using mirror reections allows imaging around corners, for example, from a glass window68. Three bounce analysis of a time-of-ight camera can recover hidden 101 planar barcodes9,10 but the technique assumes well-separated, isolated hidden patches with known correspondence between hidden patches and recorded pulses. Similar to these and other inverse light transport approaches11, we use a light source to illuminate one scene spot at a time and record the reected light aer its interaction with the scene.
We demonstrate an incoherent ultrafast imaging technique to recover three-dimensional (3D) shapes of non-line-of-sight objects using this diusely reected light. We illuminate the scene with a short pulse and use the time of ight of returning light as a means to analyse direct and scattered light from the scene. We show that the extra temporal dimension of the observations under very high temporal sampling rates makes the hidden 3D structure observable. With a single or a few isolated hidden patches, pulses recorded aer reections are distinct and can be easily used to nd 3D positions of the hidden patches. However, with multiple hidden scene points, the reected pulses may overlap in both space and time when they arrive at the detector. The loss of correspondence between 3D scene points and their contributions to the detected pulse stream is the main technical challenge. We present a computational algorithm based on backprojection to invert this process. Our main contributions are twofold. We introduce the new problem of recovering the 3D structure of a hidden object and we show that the 3D information is retained in the temporal dimension aer multi-bounce interactions between visible and occluded parts. We also present an experimental realization of the ability to recover the 3D structure of a hidden object, thereby demonstrating a 3D range camera able to look around a corner. The ability to record 3D shapes beyond the line of sight can potentially be applied in industrial inspection, endoscopy, disaster relief scenarios, or more generally, in situations where direct imaging of a scene is impossible.
ResultsImaging process. The experimental set-up is shown in Fig. 1. Our scene consists of a 40-cm high and 25-cm wide wall referred to as the diuser wall. We use an ultrafast laser and a streak camera and both are directed at this wall. As a time reference, we also direct an attenuated portion of the laser beam into the eld of view of the streak camera (Fig. 2). The target object is hidden in the scene (mannequin in Fig. 1), so that direct light paths between the object and the laser or the camera are blocked. Our goal is to produce 3D range data of the target object.
The streak camera records a streak image with one spatial dimension and one temporal dimension. We focus the camera on the dashed line segment on the diuser wall shown in Fig. 1a. We arrange the scanning laser to hit spots on the wall above or below this line segment so that single bounce light does not enter the camera. Though the target object is occluded, light from the laser beam is diusely reected by the wall, reaches the target object, is reected by multiple surface patches, and returns back to the diuser wall, where it is reected again and captured by the camera. In a traditional camera, this image would contain little or no information
Wall
W
Beam splitters
Confidence (a.u.)
Ultrafast laser
Photo-detector
672
Galvo
L
Streak camera
r4
r1
r2
r3
Object (hidden)
Intensity (a.u.)
1
s
T(ns)
Occluder
0 0 X (pixels)
Figure 1 | Experimental set-up. (a) The capture process: we capture a series of images by sequentially illuminating a single spot on the wallwith a pulsed laser and recording an image of the dashed line segmenton the wall with a streak camera. The laser pulse travels a distance r1 to strike the wall at a point L; some of the diffusely scattered light strikes the hidden object (for example at s after travelling a distance r2), returns to the wall (for example at w, after travelling over r3) and is collected by the camera after travelling the nal distance r4 from w to the camera centreof projection. The position of the laser beam on the wall is changed by a set of galvanometer-actuated mirrors. (b) An example of streak images sequentially collected. Intensities are normalized against a calibration signal. Red corresponds to the maximum, blue to the minimum intensities. (c) The 2D projected view of the 3D shape of the hidden object, as recovered by the reconstruction algorithm. Here the same colour map corresponds to backprojected ltered intensities or condence values of nding an object surface at the corresponding voxel.
Intensity (a.u.)
1
0
t(ns)
0
X (pixel)
672
Figure 2 | Streak image with calibration spot. The calibration spot in a streak image (highlighted with an arrow). The calibration spot is created by an attenuated beam split off the laser beam that strikes the wall in the eld of view of the camera. It allows monitoring of the long-term stability of the system and calibration for drifts in timing synchronization.
NATURE COMMUNICATIONS | 3:745 | DOI: 10.1038/ncomms1747 | www.nature.com/naturecommunications
2012 Macmillan Publishers Limited. All rights reserved.
NATURE COMMUNICATIONS | DOI: 10.1038/ncomms1747
ARTICLE
Intensity (a.u.)
Intensity (a.u.)
Intensity (a.u.)
p q
r
p
q
r
Laser
Object
Camera
Backprojected intensity (a.u.)Backprojected intensity (a.u.)
Backprojected intensity (a.u.) Confidence (a.u.)
p
q
r
Figure 3 | Reconstruction algorithm. An illustrative example of geometric reconstruction using streak camera images. (a) Data capture. The object to be recovered consists of a 2 cm2 cm size square white patch beyond the line of sight (that is, hidden). The patch is mounted in the scene anddata is collected for different laser positions. The captured streak images corresponding to three different laser positions are displayed in the top row. Shapes and timings of the recorded response vary with laser positions and encode the position and shape of the hidden patch. (b) Contributing voxelsin Cartesian space. For recovery of hidden position, consider the choices of contributing locations. The possible locations in Cartesian space that could have contributed intensity to the streak image pixels p, q, r are the ellipses p, q, r (ellipsoids in 3D). For illustration, these three ellipse sections are also shown in (a) bottom left in Cartesian coordinates. If there is a single world point contributing intensity to all 3 pixels, the corresponding ellipses intersect, as is the case here. The white bar corresponds to 2 cm in all sub-gures. (c) Backprojection and heatmap. We use a backprojection algorithm that nds overlayed ellipses corresponding to all pixels, Here we show summation of elliptical curves from all pixels in the rst streak image. (d) Backprojection using all pixels in a set of 59 streak images. (e) Filtering. After ltering with a second derivative, the patch location and 2-cm lateral size are recovered.
about the occluded target object (Supplementary Figs S5 and S6; Supplementary Methods).
In our experimental set-up, the laser emits 50-fs long pulses. The camera digitizes information in time intervals of 2 ps. We assume the geometry of the directly visible part of the set-up is known. Hence, the only unknown distances in the path of the laser pulses are those from the diuser wall to the dierent points on the occluded target object and back (paths r2 and r3 in Fig. 1). The 3D geometry of the occluded target is thus encoded in the streak images acquired by the camera and decoded using our reconstruction algorithm. The recorded streak images lack correspondence information, that
is, we do not know which pulses received by the camera came from which surface point on the target object. Hence, a straightforward triangulation or trilateration to determine the hidden geometry is not possible.
Consider a simple scenario with a small hidden patch as illustrated in Fig. 3a. It provides intuition on how the geometry and location of the target object are encoded in the streak images. The reected spherical pulse propagating from the hidden patch arrives at the points on the diuser wall with dierent time delays and creates a hyperbolic curve in the space-time streak image. We scan and successively change the position of the laser spot on the diuser
NATURE COMMUNICATIONS | 3:745 | DOI: 10.1038/ncomms1747 | www.nature.com/naturecommunications
2012 Macmillan Publishers Limited. All rights reserved.
ARTICLE
NATURE COMMUNICATIONS | DOI: 10.1038/ncomms1747
Intensity (a.u.) Backprojected intensity (a.u.)
Confidence (a.u.)
Depth (a.u.)
Near Far
Figure 4 | Complex object reconstruction. (a) Photo of the object. The mannequin is ~20 cm tall and is placed about 25 cm from the diffuser wall. (b) Nine of the sixty raw streak images. (c) Heatmap. Visualization of the heatmap after backprojection. The maximum value along the z direction for each xy coordinate in Cartesian space. The hidden shape is barely discernible. (d) Filtering. The second derivative of the heatmap along depth (z) projected on the xy plane reveals the hidden shape contour. (e) Depth map. Colour-encoded depth (distance from the diffuser wall) shows the left leg and right arm closer in depth compared with the torso and other leg and arm. (f) Condence map. A rendered point cloud of condence values after soft threshold. Images (g,h) show the object from different viewpoints after application of a volumetric blurring lter. (i) The stop-motion animation frames from multiple poses to demonstrate reproducability. See Supplementary Movie 1 for an animation. Shadows and the ground plane in images (fi) have been added to aid visualization.
wall. The shape and position of the recorded hyperbolic curve varies accordingly. Each pixel in a streak image corresponds to a nite area on the wall and a 2-ps time interval, a discretized space-time bin. However, the eective time resolution of the system is 15 ps owing to a nite temporal-point spread function of the camera. The detailed description of image formation is included in the Supplementary Methods.
The inverse process to recover the position of the small hidden patch from the streak images is illustrated in Fig. 3be. Consider three pixels p, q and r in the streak image at which non-zero light intensity is measured (Fig. 3a). The possible locations in the world that could have contributed to a given pixel lie on an ellipsoid in
Cartesian space (see also Supplementary Fig. S1). The foci of this ellipsoid are the laser spot on the diuser wall and the point on the wall observed by that pixel. For illustration, we draw only a 2D slice of the ellipsoid, that is, an ellipse, in Fig. 3b. The individual ellipses from each of the three pixels p, q and r intersect at a single point. In the absence of noise, the intersection of three ellipses uniquely determines the location of the hidden surface patch that contributed intensity to the three camera pixels. In practice, we lack correspondence, that is, we do not know whether or not light detected at two pixels came from the same 3D surface point.
Therefore, we discretize the Cartesian space into voxels and compute the likelihood of the voxel being on a hidden surface. For each
NATURE COMMUNICATIONS | 3:745 | DOI: 10.1038/ncomms1747 | www.nature.com/naturecommunications
2012 Macmillan Publishers Limited. All rights reserved.
NATURE COMMUNICATIONS | DOI: 10.1038/ncomms1747
ARTICLE
Intensity (a.u.)
Backprojected intensity (a.u.) Confidence (a.u.)
e
y
z
253 mm
245 mm
222 mm
x
Figure 5 | Depth in reconstructions. Demonstration of the depth and lateral resolution. (a) The hidden objects to be recovered are three letters, I, T, I at varying depths. The I is 1.5 cm wide and all letters are 8.2 cm high. (b) 9 of 60 images collected by the streak camera. (c) Projection of the heatmap created by the back projection algorithm, on the xy plane. (d) Filtering after computing second derivative along depth (z). The colour in these images represents the condence of nding an object at the pixel position. (e) A rendering of the reconstructed 3D shape. Depth is colour coded and semi-transparent planes are inserted to indicate the ground truth. The depth axis is scaled to aid visualization of the depth resolution.
voxel, we determine all streak image pixels that could potentially have received contributions of this voxel based on the time-of-ight r1 + r2 + r3 + r4 and sum up the measured intensity values in these pixels. In eect, we let each pixel vote for all points on the corresponding ellipsoid. The signal energy contributed by each pixel is amplied by a factor to compensate for the distance attenuation. If the distance attenuation factor were not accounted for, the scene points that are far away from the wall would be attenuated by a factor of (r2r3)2 and would be lost during the reconstruction. Therefore, we amplify the contribution of each pixel to a particular voxel by a factor of (r2r3) before backprojection. Reconstruction quality depends weakly on the value of . We experimented with various
values of and found that = 1 is a good choice for reduced
computation time. This process of computing likelihood by summing up weighted intensities is called backprojection12. We call the resulting 3D scalar function on voxels a heatmap.
The summation of weighted intensities from all pixels in a single- streak image creates an approximate heatmap for the target patch (Fig. 3c). Repeating the process for many laser positions on the diffuser wall, and using pixels from the corresponding streak images provides a better approximation (Fig. 3d and Supplementary Fig. S2). In practice, we use ~60 laser positions. Traditional backprojection requires a high-pass ltering step. We use the second derivative of the data along the z direction of the voxel grid and approximately perpendicular to the wall as an eective lter and recover the hidden surface patch in Fig. 3e. Because values
NATURE COMMUNICATIONS | 3:745 | DOI: 10.1038/ncomms1747 | www.nature.com/naturecommunications
2012 Macmillan Publishers Limited. All rights reserved.
ARTICLE
NATURE COMMUNICATIONS | DOI: 10.1038/ncomms1747
at the voxels in the heatmap are the result of summing a large number of streak image pixels, the heatmap contains low noise and the noise amplication associated with a second-derivative lter is acceptable.
Algorithm. The rst step of our imaging algorithm is data acquisition. We sequentially illuminate a single spot on the diuser wall with a pulsed laser and record an image of the line segment of the wall with a streak camera. Then, we estimate an oriented bounding box for the working volume to set up a voxel grid in Cartesian space (see Methods). In the backprojection, for each voxel, we record the summation of weighted intensities of all streak image pixels that could potentially have received contributions of this voxel based on the time of ight. We store the resulting 3D heatmap of voxels. The backprojection is followed by ltering. We compute a second derivative of the heatmap along the direction of the voxel grid facing away from the wall. In an optional post processing step, we compute a condence value for each voxel by computing local contrast with respect to the voxel neighbourhood in the ltered heatmap. To compute contrast, we divide each voxel heatmap value by the maximum in the local neighbourhood. For better visualization, we apply a so threshold on the voxel condence value.
We estimate the oriented bounding box of the object in the second step by running the above algorithm at low spatial target resolution and with down-sampled input data. Details of the reconstruction process and the algorithm can be found in the Methods as well as in the Supplementary Methods.
Reconstructions. We show results of the 3D reconstruction for multipart objects in Figs 4 and 5. The mannequin in Fig. 4 contains nonplanar surfaces with variations in depth and occlusions. We accurately recover all major geometrical features of the object. Figure 4i shows the reconstruction of the same object in slightly dierent poses to demonstrate the reproducibility and stability of the method as well as the consistency in the captured data. The sporadic inaccuracies in the reconstruction are consistent across poses and are conned to the same 3D locations. The stop-motion animation in Supplementary Movie 1 shows the local nature of the missing or phantom voxels more clearly; another stop motion reconstruction is shown in Supplementary Fig. S4. Hence, the persistent inaccuracies are not due to signal noise or random measurement errors. This is promising as the voxel condence errors are primarily due to limitations in the reconstruction algorithm and instrument calibration. These limitations can be overcome with more sophistication. Our current method is limited to diffuse reection from near-Lambertian opaque surfaces. Parts of the object that are occluded from the diuser wall or facing away from it are not reconstructed.
Figure 5 shows a reconstruction of multiple planar objects at dierent unknown depths. The object planes and boundaries are reproduced accurately to demonstrate depth and lateral resolution. For further investigation of resolution see Supplementary Figs S7 and S8. The reconstruction is aected by several factors such as calibration, sensor noise, scene size and time resolution. Below, we consider them individually.
The sources of calibration errors are lens distortions on the streak camera that lead to a warping of the collected streak image, measurement inaccuracies in the visible geometry, and measurement inaccuracies of the centre of projection of the camera and the origin of the laser. For larger scenes, the impact of static calibration errors would be reduced.
The sensor introduces intensity noise and timing uncertainty, that is, jitter. The reconstruction of 3D shapes is more dependent on the accuracy of the time of arrival than the signal-to-noise ratio (SNR) in received intensity. Jitter correction, as described in the Methods, is essential, but does not remove all uncertainties. Improving the
SNR is desirable because it yields faster capture times. Similar to many commercial systems, for example, LIDAR, the SNR could be signicantly improved by using an amplied laser with more energetic pulses and a repetition rate in the kilohertz range and a triggered camera. The overall light power would not change, but fewer measurements for light collection could signicantly reduce signal independent noise such as background and shot noise.
We could increase the scale of the system for larger distances and bigger target objects. By using a longer pulse, with proportionally reduced target resolution and increased aperture size, one could build systems without any change in the ratio of received and emitted energy, that is, the link budget. When the distance r2 between diuser wall and the hidden object (Fig. 1) is increased without increasing the size of the object, the signal strength drops dramatically (1/(r2r3)2) and the size of the hidden scene is therefore limited. A conguration where laser and camera are very far from the rest of the scene is, however, plausible. A loss in received energy can be reduced in two ways. The laser beam can be kept collimated over relatively long distances and the aperture size of the camera can be increased to counterbalance a larger distance between camera and diuser wall.
The timing resolution, along with spatial diversity in the positions of spots illuminated and viewed by the laser, and the camera aects the resolution of 3D reconstructions. Extra factors include the position of the voxel in Cartesian space and the overall scene complexity. The performance evaluation subsection of the Supplementary Methods describes depth and lateral resolution. In our system, translation along the direction perpendicular to the diuser wall can be resolved with a resolution of 400 mbetter than the full width half maximum time resolution of the imaging system. Lateral resolution in a plane parallel to the wall is lower and is limited to 0.51 cm depending on proximity to the wall.
Discussion
This papers goals are twofold: to introduce the new challenging problem of recovering the 3D shape of a hidden object and to demonstrate the results using a novel co-design of an electro-optic hardware platform and a reconstruction algorithm. Designing and implementing a prototype for a specic application will provide further, more specic data about the performance of our approach in real-world scenarios. We have demonstrated the 3D imaging of a nontrivial hidden 3D geometry from scattered light in free space. We compensate for the loss of information in the spatial light distribution caused by the scattering process by capturing ultrafast time-of-ight information.
Our reconstruction method assumes that light is only reected once by a discrete surface on the hidden object without inter-reections within the object and without subsurface scattering. We further assume that light travels in a straight line between reections. Light that does not follow these assumptions will appear as time-delayed background in our heatmap and will complicate, but not necessarily prevent, reconstruction.
The application of imaging beyond the line of sight is of interest for sensing in hazardous environments, such as inside machinery with moving parts, for monitoring highly contaminated areas such as the sites of chemical or radioactive leaks where even robots cannot operate or need to be discarded aer use13. Disaster response and search and rescue planning, as well as autonomous robot navigation, can benet from the ability to obtain complete information about the scene quickly14,15.
A promising theoretical direction is in inference and inversion techniques that exploit scene priors, sparsity, rank, meaningful transforms, and achieve bounded approximations. Adaptive sampling can decide the next-best laser direction based on a current estimate of the 3D shape. Further analysis will include coded sampling using compressive techniques and noise models for SNR and
NATURE COMMUNICATIONS | 3:745 | DOI: 10.1038/ncomms1747 | www.nature.com/naturecommunications
2012 Macmillan Publishers Limited. All rights reserved.
NATURE COMMUNICATIONS | DOI: 10.1038/ncomms1747
ARTICLE
eective bandwidth. Our current demonstration assumes friendly reectance and planarity of the diuse wall.
The reconstruction of an image from diusely scattered light is of interest in a variety of elds. Change in spatial light distribution due to the propagation through a turbid medium is in principle reversible16 and allows imaging through turbid media via computational imaging techniques1719. Careful modulation of light can shape or focus pulses in space and time inside a scattering medium20,21. Images of objects behind a diuse screen, such as a shower curtain, can be recovered by exploiting the spatial frequency domain properties of direct and global components of scattered light in free space22. Our treatment of scattering is dierent but could be combined with many of these approaches.
In the future, emerging integrated solid-state lasers, new sensors and nonlinear optics should provide practical and more sensitive imaging devices. Beyond 3D shape, new techniques should allow us to recover reectance, refraction and scattering properties and achieve wavelength-resolved spectroscopy beyond the line of sight. The formulation could also be extended to shorter wavelengths (for example, X-rays) or to ultrasound and sonar frequencies. The new goal of hidden 3D shape recovery may inspire new research in the design of future ultrafast imaging systems and novel algorithms for hidden scene reconstruction.
Methods
Capture set-up. The light source is a Kerr lens mode-locked Ti:sapphire laser. It delivers pulses of about 50 fs length at a repetition rate of 75 MHz. Dispersion in the optical path of the pulse does not stretch the pulse beyond the resolution of the camera of 2 ps and therefore can be neglected. The laser wavelength is centred at 795 nm. The main laser beam is focused on the diuser wall with a 1-m focal length lens. The spot created on the wall is about 1 mm in diameter and is scanned across the diuser wall through a system of two galvanometer-actuated mirrors.
A small portion of the laser beam is split o with a glass plate and is used to synchronize the laser and streak camera as shown in Fig. 1. The diuser wall is placed 62 cm from the camera. The mannequin object (Fig. 4) is placed at a distance of about 25 cm to the wall, the letters (Fig. 5) are placed at 22.225.3 cm.
For time-jitter correction, another portion of the beam is split o, attenuated and directed at the wall as the calibration spot. The calibration spot is in the direct eld of view of the camera and can be seen in Fig. 2. The calibration spot servesas a time and intensity reference to compensate for dris in the synchronization between laser and camera as well as changes in laser-output power. It also helps in detecting occasional shis in the laser direction due to, for example, beam pointing instabilities in the laser. If a positional shi is detected, the data is discarded and the system is re-calibrated. The streak cameras photocathode tube, much like an oscilloscope, has time decayed burn out and local gain variations. We use a reference background photo to divide and compensate.
The camera is a Hamamatsu C5680 streak camera that captures one spatial dimension, that is, a line segment in the scene, with an eective time resolution of 15 ps and a quantum efficiency of about 10%. The position and viewing direction of the camera are xed. The diuser wall is covered with Edmund Optics NT83 diuse white paint.
Reconstruction technique. We use a set of Matlab routines to implement the backprojection-based reconstruction. Geometry information about the visiblepart of the scene, namely diuser wall, could be collected using our time-of-ight system. Reconstructing 3D geometry of a visible scene using time-of-ight data is well known2. We omit this step and concentrate on the reconstruction of the hidden geometry. We use a FARO Gauge digitizer arm to measure the geometry of the visible scene and also to gather data about a sparse set of points on hidden objects for comparative verication. The digitizer arm data is used as ground truth for later independent verication of the position and shape of hidden objects as shown via transparent planes in Fig. 4e. Aer calibration, we treat the camera and laser as a rigid pair with known intrinsic and extrinsic parameters23.
We estimate the oriented bounding box around the hidden object using a lower resolution reconstruction. We reduce the spatial resolution to 8 mm per voxel, and downsample the input data by a factor of 40. We can scan a 40 cm40 cm40 cm volume spanning the space in front of the wall in 24 s to determine the bounding box of a region of interest. The ner voxel grid resolution is 1.7 mm in each dimension. We can use the coarse reconstruction obtained to set up a ner grid within this bounding box. Alternatively, we can set an optimized bounding box from the collected ground truth. To minimize reconstruction time, we used this second method in most of the published reconstructions. We conrmed that apart from the reconstruction time and digitization artefacts, both methods produce the same results. We compute the principal axis of this low-resolution approximation, and orient the ne voxel grid with these axes.
In the post-processing step, we use a common approach to improve the surface visualization. We estimate the local contrast and apply a so threshold. The condence value for a voxel is V = tanh(20(V V0))V/mloc, where V is the original voxel value in ltered heatmap and mloc is a local maximum computed in a 202020 voxel sliding window around the voxel under consideration. Division by mloc normalizes for local contrast. The value V0 is a global threshold and set to 0.3 times the global maximum of the ltered heatmap. The tanh function achieves a so threshold.
System SNR. The laser emits a pulse every 13.3 ns (75 MHz) and consequently the reected signal repeats at the same rate. We average 7.5 million such 13.3 ns windows in a 100-ms exposure time on our streak tube readout camera. Weadd 50200 such images to minimize noise from the readout camera. The light returned from a single hidden patch is attenuated in the second, third and fourth path segments. In our set-ups, this attenuation factor is 10 8. Attenuation in the fourth path segment can be partially counteracted by increasing the camera aperture.
Choice of laser positions on the wall. Recall that we direct the laser to various positions on the diuser wall and capture one streak image for each position. The position of a hidden point s (Fig. 1) is determined with highest condence along the normal N to an ellipsoid through s with foci at the laser spot, L, and the wall point, w. Large angular diversity through a wide range of angles for N for all such pairs to create baselines is important. N is the angle bisector of Lsw.
The location and spacing of the laser positions on the wall can have a big impact on reconstruction performance. To obtain good results, one should choose the laser positions so as to provide good angular diversity. We use 60 laser positions in 35 lines perpendicular to the line on the wall observed by our one-dimensional streak camera. This conguration yielded signicantly better results than putting the laser positions on few lines parallel to the camera line.
Scaling the system. Scaling up the distances in the scene is challenging because higher resolution and larger distances lead to disproportionately less light being transferred through the scene. A less challenging task may be to scale the entire experiment including the hidden object, the pulse length, the diuser wall and the camera aperture. The reduction in resolution to be expected in this scaling should be equal to the increase in size of the hidden object.
To understand this, consider a hidden square patch in the scene. To resolve it, we require discernible light to be reected back from that patch aer reectionsor bounces o other patches. Excluding the collimated laser beam, there are three paths, as described earlier. For each path, light is attenuated by approximately d2/(2r2), where r is the distance between the source and the destination patch and d is the side length of the destination patch. For the fourth segment, the destination patch is the camera aperture and d denotes the size of this aperture. If r and d are scaled together for any path, the contributed energy from the source patch to the destination patch does not change. This may allow us to scale the overall systemto larger scenes without a prohibitively drastic change in performance. However, increasing the aperture size is only possible to a certain extent.
Non-Lambertian surfaces. Our reconstruction method is well-suited for Lambertian reectance of surfaces. Our method is also robust for near-Lambertian surfaces, for example, surfaces with a large diuse component, and they are implicitly handled in our current reconstruction algorithm. The surface reectance prole only aects the relative weight of the backprojection ellipses and not their shapes. The shape is dictated by the time-of-ight, which is independent of the reectance distribution.
Surfaces that are highly specular, retroreective or have a low reectance make the hidden shape reconstruction challenging. Highly specular, mirror-like and retroreective surfaces limit the regions illuminated by the subsequent bounces and may not reect enough energy back to the camera. They also could cause dynamic range problems. Subsurface scattering or extra inter-reections extend the fall time in reected time prole of a pulse. But the onset due to reection from the rst surface is maintained in the time prole and hence the time delayed reections appear as background noise in our reconstruction. Absorbing low-reectance black materials reduce the SNR but the eect is minor compared with the squared attenuation over distances.
Although near-Lambertian surfaces are very common in the proposed application areas, reconstruction in the presence of varying reectance materials is an interesting future research topic.
References
1. Kamerman, G. W. Laser Radar [M]. Chapter 1 of Active Electro-Optical System, Vol. 6, The Infrared and Electro-Optical System Handbook (1993).
2. Busck, J. & Heiselberg, H. Gated viewing and high-accuracy three-dimensional laser radar. Appl. Optics 43, 47054710 (2004).
3. Wang, L., Ho, P. P., Liu, C., Zhang, G. & Alfano, R. R. Ballistic 2-D imaging through scattering walls using an ultrafast optical Kerr gate. Science 253, 769771 (1991).
NATURE COMMUNICATIONS | 3:745 | DOI: 10.1038/ncomms1747 | www.nature.com/naturecommunications
2012 Macmillan Publishers Limited. All rights reserved.
ARTICLE
NATURE COMMUNICATIONS | DOI: 10.1038/ncomms1747
4. Xia, H. & Zhang, C. Ultrafast ranging lidar based on real-time Fourier transformation. Opt. Lett. 34, 21082110 (2009).
5. Sen, P. et al. Dual photography. Proc. of ACM SIGGRAPH 24, 745755 (2005).6. Repasi, E. et al. Advanced short-wavelength infrared range-gated imaging for ground applications in monostatic and bistatic congurations. Appl. Optics 48, 59565969 (2009).
7. Sume, A. et al. Radar detection of moving objects around corners. Proc. of SPIE 7308, 73080V73080V (2009).
8. Chakraborty, B. et al. Multipath exploitation with adaptive waveform design for tracking in urban terrain. IEEE ICASSP Conference 38943897 (2010).9. Raskar, R. & Davis, J. 5d time-light transport matrix: what can we reasonabout scene properties? MIT Technical Report (2008). URL:http://hdl.handle. net/1721.1/67888.
10. Kirmani, D., Hutchison, T., Davis, J. & Raskar, R. Looking around the corner using transient imaging. Proc. IEEE ICCV 159166 (2009).
11. Seitz, S. M., Matsushita, Y. & Kutulakos, K. N. A theory of inverse light transport. Proc. IEEE ICCV 2, 14401447 (2005).
12. Kak, A. C. & Slaney, M. Principles of Computerized Tomographic Imaging, Chapter 3. (IEEE Service Center, 2001).
13. Blackmon, T. T. et al. Virtual reality mapping system for chernobyl accident site assessment. Proc. SPIE 3644, 338345 (1999).
14. Burke, J. L., Murphy, R. R., Coovert, M. D. & Riddle, D. L. Moonlight in miami: a eld study of human-robot interaction in the context of an urban search and rescue disaster response training exercise. Hum.-Comput. Interact. 19, 85116 (2004).
15. Ng, T. C. et al. Vehicle following with obstacle avoidance capabilities in natural environments. Robotics and Automation, 2004. Proc. ICRA 04. 2004 IEEE International Conference on 5, 42834288 (2004).
16. Feld, M. S., Yaqoob, Z., Psaltis, D. & Yang, C. Optical phase conjugation for turbidity suppression in biological samples. Nature Photon. 2, 110115 (2008).
17. Dylov, D. V. & Fleischer, J. W. Nonlinear self-ltering of noisy images via dynamical stochastic resonance. Nature Photon. 4, 323328 (2010).
18. Popo, S., Lerosey, G., Fink, M., Boccara, A. C. & Gigan, S. Image transmission through an opaque material. Nat. Commun. 1, 81 (2010).
19. Vellekoop, I. M., Lagendijk, A. & Mosk, A. P. Exploiting disorder for perfect focusing. Nature Photon. 4, 320322 (2010).
20. Choi, Y. et al. Overcoming the diraction limit using multiple light scattering in a highly disordered medium. Phys. Rev. Lett. 107, 023902 (2011).
21. Katz, O., Small, E., Bromberg, Y. & Silberberg, Y. Focusing and compression of ultrashort pulses through scattering media. Nature Photon. 5, 372377 (2011).
22. Nayar, S. K., Krishnan, G., Grossberg, M. D. & Raskar, R. Fast separation of direct and global components of a scene using high frequency illumination. ACM Trans. Graph. 25, 935944 (2006).
23. Forsyth, D. & Ponce, J. Computer Vision, A Modern Approach (Prentice Hall, 2002).
Acknowledgements
This work was funded by the Media Lab Consortium Members, DARPA through the DARPA YFA grant, and the Institute for Soldier Nanotechnologies and U.S. Army Research Office under contract W911NF-07-D-0004. A. Veeraraghavan was supported by the NSF IIS Award #1116718. We would like to thank Amy Fritz for her help with acquiring and testing data and Chinmaya Joshi for his help with the preparation of the gures.
Author contributions
The method was conceived by A. Velten and R.R., the experiments were designed by A.Velten, M.G.B. and R.R. and performed by A.Velten. The algorithm was conceived by A.Velten, T.W., O.G., and R.R. and implemented and optimized by A.Velten, T.W., O.G., A.Veeraraghavan, and R.R. All authors took part in writing the paper.
Additional information
Supplementary Information accompanies this paper at http://www.nature.com/ naturecommunications
Competing nancial interests: The authors declare no competing nancial interests.
Reprints and permission information is available online at http://npg.nature.com/ reprintsandpermissions/
How to cite this article: Velten, A. et al. Recovering three-dimensional shape around a corner using ultrafast time-of-ight imaging. Nat. Commun. 3:745 doi: 10.1038/ncomms1747 (2012).
NATURE COMMUNICATIONS | 3:745 | DOI: 10.1038/ncomms1747 | www.nature.com/naturecommunications
2012 Macmillan Publishers Limited. All rights reserved.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright Nature Publishing Group Mar 2012
Abstract
The recovery of objects obscured by scattering is an important goal in imaging and has been approached by exploiting, for example, coherence properties, ballistic photons or penetrating wavelengths. Common methods use scattered light transmitted through an occluding material, although these fail if the occluder is opaque. Light is scattered not only by transmission through objects, but also by multiple reflection from diffuse surfaces in a scene. This reflected light contains information about the scene that becomes mixed by the diffuse reflections before reaching the image sensor. This mixing is difficult to decode using traditional cameras. Here we report the combination of a time-of-flight technique and computational reconstruction algorithms to untangle image information mixed by diffuse reflection. We demonstrate a three-dimensional range camera able to look around a corner using diffusely reflected light that achieves sub-millimetre depth precision and centimetre lateral precision over 40 cm×40 cm×40 cm of hidden space.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer