virtual light transport matrices for non-line-of-sight imaging

10
Virtual light transport matrices for non-line-of-sight imaging Julio Marco 1 Adrian Jarabo 1 Ji Hyun Nam 2 Xiaochun Liu 2 Miguel ´ Angel Cosculluela 1 Andreas Velten 2 Diego Gutierrez 1 1 Universidad de Zaragoza, I3A 2 University of Wisconsin – Madison Abstract The light transport matrix (LTM) is an instrumental tool in line-of-sight (LOS) imaging, describing how light interacts with the scene and enabling applications such as relighting or separation of illumination components. We introduce a framework to estimate the LTM of non- line-of-sight (NLOS) scenarios, coupling recent virtual for- ward light propagation models for NLOS imaging with the LOS light transport equation. We design computational projector-camera setups, and use these virtual imaging sys- tems to estimate the transport matrix of hidden scenes. We introduce the specific illumination functions to compute the different elements of the matrix, overcoming the challeng- ing wide-aperture conditions of NLOS setups. Our NLOS light transport matrix allows us to (re)illuminate specific locations of a hidden scene, and separate direct, first-order indirect, and higher-order indirect illumination of complex cluttered hidden scenes, similar to existing LOS techniques. 1. Introduction The light transport matrix (LTM) [26] is a fundamen- tal tool to understand how a scene transports incident light, describing the linear response of the scene for a given illu- mination. It has become the cornerstone of many applica- tions such as dual photography [36], image-based relight- ing [30, 41], acquisition of material properties [5, 35], sep- aration of global illumination components [25, 32], or ro- bust depth estimation [29]. To acquire the LTM, a camera- projector setup is used, capturing the scene under varying coded illumination conditions. On the other hand, recent works on non-line-of-sight (NLOS) imaging have shown the potential to signifi- cantly advance many fields including medical imaging, au- tonomous driving, rescue operations, or defense and secu- rity, to name a few. The key idea is to leverage the scattered radiance on a secondary relay surface to infer information about the hidden scene [39, 10, 23]. However, NLOS imag- Hidden scene Direct (volume) Direct (front) Indirect (front) A B C A B C 2.0m - 2.3m 1.1m - 1.4m 0.1m - 0.4m A B C 0.0 3.2 0.0 0.6 0.0 0.25 0.0 0.9 Figure 1: We introduce a framework to compute virtual light transport matrices (LTM) of NLOS scenes. Probing the virtual LTM of a hidden scene allows us to extract its di- rect illumination (first row), the indirect components when illuminating a single point in the scene (second row, each point corresponds to a column in the LTM), or decompos- ing near-, middle-, and far-field indirect components (third row) when illuminating specific points in the hidden scene. The insets show the probed elements from the virtual LTM for each image. ing is still on its infancy, and many well-established capabil- ities of traditional line-of-sight (LOS) imaging modalities cannot yet be applied in NLOS conditions. In this work we take steps towards bridging this gap be- tween LOS and NLOS imaging modalities, introducing a computational framework to obtain the LTM of a hidden scene. In particular, we build on the recent wave-based phasor fields framework [21], which poses NLOS as a for- ward wave transport problem, creating virtual (computa- 2440

Upload: others

Post on 31-Dec-2021

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

Virtual light transport matrices for non-line-of-sight imaging

Julio Marco1 Adrian Jarabo1 Ji Hyun Nam2 Xiaochun Liu2

Miguel Angel Cosculluela1 Andreas Velten2 Diego Gutierrez1

1Universidad de Zaragoza, I3A 2University of Wisconsin – Madison

Abstract

The light transport matrix (LTM) is an instrumentaltool in line-of-sight (LOS) imaging, describing how lightinteracts with the scene and enabling applications suchas relighting or separation of illumination components.We introduce a framework to estimate the LTM of non-line-of-sight (NLOS) scenarios, coupling recent virtual for-ward light propagation models for NLOS imaging with theLOS light transport equation. We design computationalprojector-camera setups, and use these virtual imaging sys-tems to estimate the transport matrix of hidden scenes. Weintroduce the specific illumination functions to compute thedifferent elements of the matrix, overcoming the challeng-ing wide-aperture conditions of NLOS setups. Our NLOSlight transport matrix allows us to (re)illuminate specificlocations of a hidden scene, and separate direct, first-orderindirect, and higher-order indirect illumination of complexcluttered hidden scenes, similar to existing LOS techniques.

1. Introduction

The light transport matrix (LTM) [26] is a fundamen-tal tool to understand how a scene transports incident light,describing the linear response of the scene for a given illu-mination. It has become the cornerstone of many applica-tions such as dual photography [36], image-based relight-ing [30, 41], acquisition of material properties [5, 35], sep-aration of global illumination components [25, 32], or ro-bust depth estimation [29]. To acquire the LTM, a camera-projector setup is used, capturing the scene under varyingcoded illumination conditions.

On the other hand, recent works on non-line-of-sight(NLOS) imaging have shown the potential to signifi-cantly advance many fields including medical imaging, au-tonomous driving, rescue operations, or defense and secu-rity, to name a few. The key idea is to leverage the scatteredradiance on a secondary relay surface to infer informationabout the hidden scene [39, 10, 23]. However, NLOS imag-

Hidden scene Direct (volume) Direct (front)

Indirect (front)

A B C

A B C

2.0m - 2.3m1.1m - 1.4m0.1m - 0.4mA B C

0.0 3.2

0.0 0.60.0 0.250.0 0.9

Figure 1: We introduce a framework to compute virtuallight transport matrices (LTM) of NLOS scenes. Probingthe virtual LTM of a hidden scene allows us to extract its di-rect illumination (first row), the indirect components whenilluminating a single point in the scene (second row, eachpoint corresponds to a column in the LTM), or decompos-ing near-, middle-, and far-field indirect components (thirdrow) when illuminating specific points in the hidden scene.The insets show the probed elements from the virtual LTMfor each image.

ing is still on its infancy, and many well-established capabil-ities of traditional line-of-sight (LOS) imaging modalitiescannot yet be applied in NLOS conditions.

In this work we take steps towards bridging this gap be-tween LOS and NLOS imaging modalities, introducing acomputational framework to obtain the LTM of a hiddenscene. In particular, we build on the recent wave-basedphasor fields framework [21], which poses NLOS as a for-ward wave transport problem, creating virtual (computa-

2440

Page 2: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

tional) light sources and cameras at the visible relay surfacefrom time-resolved measurements of the hidden scene.

While this approach would in principle allow turningNLOS imaging into a virtual LOS problem, working in theNLOS domain introduces two main challenges which arenot present in LOS settings: 1) the large baseline of capturesetups on the visible surfaces results in a very large virtualaperture, creating a very shallow depth-of-field and there-fore significant out-of-focus contribution; and 2) the resolu-tion of the virtual projector-camera is limited, with poten-tially significant cross-talk between neighbor (virtual) pix-els. As a consequence, directly applying LOS techniques tocapture the LTM of hidden scenes would lead to suboptimalresults.

We develop computational (virtual) illumination andimaging functions, leveraging the fact that undesired con-tributions of light transport in the hidden scene are coded inits NLOS virtual LTM. Then, inspired by existing works inlight transport analysis in LOS settings [25, 32], we exploitthe LTM and demonstrate illumination decomposition inNLOS scenes (see Fig. 1). In particular, we probe differentelements of the matrix, including direct and indirect illumi-nation, as well as near-, middle-, and far-field indirect lightdecomposition. This has potential applications for relight-ing, material analysis, improved geometry reconstruction,or scene understanding in general beyond the third bounce.In summary, our contributions are:

• We introduce a framework to obtain the light transportmatrix (LTM) of a hidden scene, helping bridge thegap between LOS and NLOS imaging.

• We develop the computational methods to obtain thenecessary virtual illumination and imaging functions,dealing with the main challenges on the NLOS imag-ing modality.

• We demonstrate how our formulation allows to probethe virtual LTM, allowing to separate NLOS direct andindirect illumination components in hidden scenes.

2. Related workLight transport acquisition and analysis. Several workshave focused on acquiring the linear light transport responsefrom a illuminated scene (i.e., the light transport matrix) bymeans of exhaustive capture [5], low-rank matrix approx-imations [41] and homogeneous factorization [27], opticalcomputing [30], compressed sensing [34, 37], or machinelearning [46]. In our work we propose a framework to com-pute virtual transport matrices of NLOS scenes using virtualcamera/projector systems, which could be potentially com-bined with any of these approaches for improved efficiency.

In terms of light transport analysis, Sen et al. [36] lever-aged the Helmholtz reciprocity to create a dual camera-

projector system allowing to take photographs with a singlepixel camera. Nayar et al. [25] exploited the structure of thelinear light transport operator to disambiguate between di-rect and indirect illumination. Mukaigawa et al. [24] gener-alized this work to scattering media. Later, Nayar et al.’s ap-proach was improved by O’Toole and Kutulakos [32], whointroduced advanced probing codes in the illumination, al-lowing to disambiguate between high- and low-frequencyglobal illumination. Follow-up work [29] included the tem-poral domain, allowing high-quality depth reconstructionusing time-of-flight cameras. Our paper builds on theseworks, but shifts its domain from visible scenes to theregime of NLOS scenes, allowing to obtain virtual lighttransport matrices of occluded scenes.

Impulse NLOS imaging. Transient imaging methodsleverage information encoded in time-resolved light trans-port using ultra-fast illumination capture systems [40, 9, 6]in applications such as light-in-flight videos [40, 9, 28],bare-sensor imaging [43], or seeing through turbid me-dia [11, 44]. In the following we discuss methods relatedto NLOS imaging based on transient light transport, and re-fer to the reader to [15] for a broader overview.

This line of work follows the approach proposed by Kir-mani et al. [16] and demonstrated experimentally by Veltenand colleagues [39]: very short laser pulses illuminate a vis-ible surface facing the hidden scene, then the scattered lightreflected back onto such visible surface is captured. Thehidden scene is then reconstructed using backprojection al-gorithms and heuristic filters [39, 17, 2, 4], or inverting sim-plified light transport models [31, 1, 33, 45, 10, 47, 13].These methods have usually been demonstrated in simpleisolated scenes with little indirect illumination from inter-reflections. In contrast, recent wave-based methods forNLOS imaging [21, 18, 20] removed these limitations byposing NLOS imaging as forward virtual propagation mod-els. We leverage existing formulations of forward propaga-tion [21] to compute virtual LTMs of hidden scenes usingvirtual projector/camera pairs. Different from (and com-plementary to) classic NLOS geometry reconstruction ap-proaches, our work focuses on analyzing light transport,paving the way for relighting and separation of illumina-tion components in hidden scenes. Wu et al. [42] and Guptaet al. [8] used transient imaging to disambiguate betweendirect and global light transport in a LOS scene; in con-trast, we demonstrate direct-indirect separation in non-line-of-sight scenes by estimating and probing their light trans-port matrices.

3. Background

In the following we summarize the key aspects of thelight transport matrix and the forward NLOS formulation

2441

Page 3: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

Table 1: Notation used throughout the paper.

V Voxelized space of the hidden scene.Kv Number of voxels discretizing V .Γ ⊂ V Subspace of V of non-empty regions.xa,xb,xv∈V Points in the hidden scene.L, S Laser and SPAD capture planes.Kp,Ki Number of measurements in L and S.xl∈L,xs∈S Points at planes L and S, respectively.〈x0, ...,xn〉 A (n + 1)-vertex light path from x0 to xn.

P (x, t) Broadband complex phasor at point x.Pω(x, t) Monochromatic phasor at x with frequency

ω.Lω(xa,xb, t) Monochromatic thin-lens phasor propagator

from xa to xb with frequency ω.G (t′, t) Virtual temporal gating function centered at

time t′.

T(xa,xb) Virtual transport matrix at illuminated andimaged locations xa and xb.

Mi(xa,xb) Matricial binary mask removing the contri-bution from xa to xb

H (xl,xs, t) Time-resolved impulse response function atlaser and SPAD locations xl and xs.

of phasor fields. Table 1 defines a list of the most commonsymbols and expressions used throughout the paper.

3.1. Transport matrix formulation

Light transport is a linear process that can be encodedin a transport matrix [26] relating the response of the scenecaptured by the sensor i with the illumination p as

i = Tp, (1)

where p is a Kp-sized column vector representing the lightsources, i is a Ki-sized column vector representing the ob-servations, and T is the transport matrix of size Kp × Ki

modeling the linear response of the scene. Intuitively, eachrow of T allows us to reconstruct the captured scene as il-luminated by a single light source j, represented as the j-thcomponent of vector p. Under controlled illumination, thetransport matrix T of a scene can be trivially computed byindependent activation of all light sources p in the scene.Measuring the image response of each activated light sourcepj allows us to fill the j-th row of T. More efficient meth-ods have been developed for reconstructing T from a sparseset of measurements [41, 34, 30, 27].

Probing the LTM [25, 32] allows to separate illuminationcomponents at fast frame rates using aligned arrays of emit-ters (projectors) and sensors (camera pixels), where both pand i have the same size Kp ≡ Ki. The diagonal of theresulting square matrix T represents a good approximationof direct illumination, since every image pixel is lit onlyby its co-located projector pixel. In contrast, non-diagonal

Virtual lens

(a) (b) (c)

L S

S

Virtual sensor

Virtual apertureVirtual

illumination

Figure 2: The phasor field method leverages transient mea-surements on a visible wall to propagate virtual light wavefields (the phasor field) towards a hidden scene (a), thenimage its response (b) from a virtual line-of-sight cameraperspective (c). Figure adapted from [21].

elements of the transport matrix T encode the multiply scat-tered indirect light, where a column would represent the in-direct contribution of a single emitter over the entire scene.In our work, we show how to compute the LTM on a hid-den scene, by computationally creating a virtual projector-camera pair from a set of NLOS measurements.

3.2. Phasor field NLOS imaging

Light transport as described by Eq. 1 assumes steadystate. By adding the temporal domain t to the transportmatrix, thus taking into account propagation and scatter-ing light transport delays, T becomes the linear impulse re-sponse of the scene H(t), and Eq. 1 becomes [29]

i(t) =

∫ ∞−∞

H(τ)p(t− τ)dτ

= (H ∗ p) (t). (2)

Phasor field virtual waves [21, 7, 20] transform an NLOSimaging problem into a virtual LOS one, by creating a vir-tual imaging system on a visible diffuse relay surface. Itleverages the temporal information encoded in H on the vis-ible surface to propagate a virtual wave from a virtual emit-ter to a virtual sensor, both placed at the relay wall (Fig. 2).Matrix H is captured by sequentially illuminating a set ofpoints xl ∈ L on the visible surface, and measuring theresponse at a set of points xs ∈ S on the same surface.

Phasor fields leverage the linearity and time invariance ofH (xl,xs, t) to compute the response of the hidden scene atpoints xs in a virtual sensor, for any virtual complex-valuedemission profile P (xl, t) as

P (xs, t) =

∫L

[P (xl, t) ∗H (xl,xs, t)]dxl. (3)

Then, any point xv in the hidden scene can be imaged bypropagating the field P (xs, t) with an imaging operatorΦ(·) as

I (xv, t) = Φ (P (xs, t)) . (4)

2442

Page 4: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

This imaging operator models a virtual lens and sensorsystem, and can be formulated in terms of a Rayleigh-Sommerfeld diffraction (RSD) propagator (please refer tothe original work for details [21]). Thus, when propagatingmonochromatic signals of a single frequency ω this imageformation operator Φ (Pω(xs, t)) becomes [38]

Φ (Pω(xs, t)) =

∣∣∣∣∫S

Pω(xs, t)Lω(xs,xv)

|xv − xs|dxs

∣∣∣∣2 , (5)

where L is a complex operator that changes the phase ofP at frequency ω. While Φ could represent different imag-ing operators, we assume a thin-lens model that perfectlyfocuses a hidden location xv into a virtual image plane, so

Lω(xs,xv) = e−ik|xv−xs| (6)

where k = ω/c is the wavenumber, with c the speed oflight. By combining a focused emission signal (Eq. 3) withthe imaging operator (Eq. 5) we can image a location xv ina hidden scene as in traditional LOS imaging setups, as

I (xv, t) =

∣∣∣∣∫S

∫L

[Pω(xl, t)∗H (xl,xs, t)]

Lω(xs,xv)

|xv − xs|dxldxs

∣∣∣∣2 . (7)

Moreover, since this image formation model is entirelyvirtual and obtained by computation, the emission profilecan be any desired function without hardware restrictions.

4. The Virtual Light Transport MatrixOur goal is to compute the NLOS (virtual) light transport

matrix T for a hidden scene, given the measured impulseresponse function H. More formally, the NLOS virtualLTM T is a two-dimensional matrix where each componentT(xa,xb) represents reflected light at point xb when a unitillumination focuses at point xa. In practice, we choose acoaxial configuration where T has a size of Kv ×Kv, andKv is the number of voxels discretizing the hidden scene(i.e., the number of imaged points in three-dimensions). Toachieve our goal, we leverage the transport operators of thephasor fields framework [21], and create virtual projector-camera pairs on the visible relay surface, focusing at pointsxa and xb respectively.

Challenges. Turning the visible relay wall into a virtualprojector-camera pair is not trivial. The two main chal-lenges are: 1) While LOS imaging setups use small aper-tures to limit the amount of defocus for both the projectorand the camera (Fig. 3a), the impulse response H in the hid-den scene is captured over relatively large surfaces L and Swith respect to the reconstructed scene size (Fig. 3b). When

AB

xaxb

Aperture

Focal plane

(a) Narrow aperture (LOS)

xa

ABxb

Aperture

Focal plane

(b) Wide aperture (NLOS)

Figure 3: (a) LOS projector-camera setups are capable ofsharply illuminating and imaging elements of a scene at dif-ferent depths onto the same focal plane, due to their narrowapertures. (b) The relay wall in NLOS methods behavesas a wide aperture, where focused illumination may spreadthrough large surfaces out of the focal plane (e.g. red area ofobject A onto pixel at xa); analogously, large areas of out-of-focus objects may contribute to a single imaged pixel.

using H to propagate light through a thin lens model (Eq. 6),L and S correspond to the effective apertures of light andimage propagation, respectively. These larger virtual aper-tures result in significant out-of-focus illumination and ge-ometry during the imaging process, as Fig. 3b shows. More-over, 2) the maximum resolution in the reconstructions isbounded by the capture density of the impulse function H[22]. However, increasing this density is a time-consumingprocess, limited by the particular characteristics of the sen-sors. A suboptimal resolution also affects the sharpness ofthe illumination, and thus on the effective resolution of T.

The virtual LTM. Given the challenges described above,naıvely computing the NLOS virtual LTM will include asignificant component of out-of-focus light D, due to thelarge aperture baseline. This naıve T can be expressed as

T = T + D. (8)

We are interested in obtaining T without the influence ofundesired defocused light. In our derivations we further de-compose T into its diagonal and off-diagonal elements asT = Td + Toff. For convenience, we also decompose Dinto Dd and Doff, which represent out-of-focus illuminationat diagonal and off-diagonal elements, respectively, having

T = Td + Toff + Dd + Doff. (9)

In camera-projector setups Td = Td,1 +Td,∞ is the sum ofdirect transport Td,1 plus multiply-scattered retro-reflectionand back scattering Td,∞, while Toff is the sum of all in-direct illumination bounces Toff =

∑∞k=2 Toff,k [32]. Note

that since we are in NLOS settings, the direct illuminationand indirect components in the hidden scene correspond to3rd and 4th+ bounce illumination, respectively.

2443

Page 5: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

Modeling the illumination function. In order to removethe undesirable contribution of unfocused light encoded inDd and Doff, we derive different virtual illumination func-tions Pω(xl, t) that enable the computation of diagonal andoff-diagonal elements Td and Toff Figs. 4a and 4b illus-trate this: (a) diagonal elements, where the virtual projec-tor (red) and camera (blue) simultaneously illuminate andimage the same point (voxel) of the hidden scene; (b) off-diagonal elements, where projector and camera focus ondifferent points.

To implement the resulting virtual projector-camera sys-tem we extend the parameter space of Pω(xl, t) to take intoaccount different locations of the voxelized hidden space.In its general form, we then have

Pω(xl, t, . . . ) = G (xl, t, . . . )Lω(xl, . . . )

D (xl, . . . ). (10)

The term G represents a real-valued time-dependent gatingfunction, while Lω(xl, . . . ) is a complex-valued thin-lensoperator (Eq. 6) which propagates light from points in theillumination aperture xl ∈ L to specific locations of the hid-den space at a distance D. In the following, we develop theillumination functions Pω(xl, t, . . . ) required to computethe specific elements of the virtual LTM of NLOS scenes.

4.1. Diagonal elements

In our NLOS setting, we compute the diagonal elementsTd (i.e. T(xv,xv)) by focusing both the virtual projectorand the virtual lens at the same position xv (see Fig. 4a).As discussed before (and shown in Fig. 3b), this can in-troduce significant out-of-focus illumination Dd (see x′v inFig. 4a). To avoid such contribution, we use a gating func-tion centered at the time-of-flight of the path 〈xl,xv,xs〉,thus removing the contribution of longer path lengths. Froman NLOS perspective, this gating function isolates three-bounce light paths between the laser and the SPAD. Sincethe rise of backprojection algorithms [2, 39, 3, 17], iso-lating the third-bounce illumination (also known as tem-poral focusing) has been a recurrent approach to improverobustness in NLOS reconstructions [33, 21]. Implemen-tations of this temporal focusing include explicitly select-ing temporal bins from captured transients in backprojec-tion methods [39], gating the illumination during the acqui-sition process [33], or convolving virtual propagators withGaussian pulses [21]. We follow this last option, and useunit-amplitude Gaussian gating functions G (t′, t) centeredat t with standard deviation σ so approximately 99% ofthe Gaussian covers four times the propagation wavelength.The resulting illumination function Pω(〈xl,xv,xs〉 , t) thusbecomes

Pω(〈xl,xv,xs〉 , t) = G (td, t)Lω(xl,xv)

|xl − xv|, (11)

with td = (|xs − xv|+ |xl − xv|) /c.Finally, combining the imaging process in Eq. 7 with the

illumination function in Eq. 11 we compute each diagonalelement of Td as

Id (xv, t :=0) =

∣∣∣∣∫S

∫L

[Pω(〈xl,xv,xs〉 , t)∗H (xl,xs, t)]

Lω(xs,xv)

|xv − xs|dxldxs

∣∣∣∣2 . (12)

Discussion: In practice, this links our virtual NLOS LTMwith the confocal camera proposed by Liu et al. [21]. Notethat Eq. 12 actually computes the direct illumination Td,1component of Td, while removing the contribution of Td,∞.Computing Td,1 is required for geometry estimation fromthe third bounce. Our objective, however, is to also sepa-rate the indirect component to enable NLOS light transportanalysis.

4.2. Off-diagonal elements

The off-diagonal matrix Toff models all light reflectedoff xb when focusing light on another point xa, i.e.T(xa,xb) for xa 6= xb. It represents the indirect illumi-nation of the scene [32], and in our NLOS configurationcorresponds to light paths of the form 〈xl,xa, ...,xb,xs〉.As in the case of the diagonal elements Td, direct computa-tion of the off-diagonal elements might result in significantout-of-focus contribution Doff. To minimize this, we designa second illumination function as follows.

First, we decompose Toff = Toff,2 + Toff,3−∞, whereToff,2 represents the contribution of two-bounce illumina-tion, and Toff,3−∞ the remaining higher-order bounces. Wefocus on two-bounce illumination with 4-vertex paths of theform 〈xl,xa,xb,xs〉 (Fig. 4b, left), and discard the contri-bution of higher-order scattering since it decreases expo-nentially with each bounce.

An illuminated point xa will bounce light towards apoint xb following a sub-path ~vab = xb − xa with timeof flight tab = |xb − xa|c−1. To compute the coeffi-cient Toff,2(xa,xb) of our NLOS LTM we need to 1) fo-cus direct illumination on the source point xa, 2) gate lightpaths centered at the time of flight of the 4-vertex path〈xl,xa,xb,xs〉, and 3) image point xb. Since the positionsof all four vertices are known, paths of this form can beisolated by using a narrow temporal gating function as (weremove the path-dependence for clarity)

ti,4 = (|xa − xl|+ |xb − xa|+ |xs − xb|) c−1.

For gating, we use the same unit-amplitude Gaussian gatingfunction as in the diagonal elements, but centered at ti,4 asG (ti,4, t). A gating function for higher-order bounces is de-scribed in Sec. A. The illumination and imaging operators

2444

Page 6: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

xv

Aperturexlxs

T(xv,xv)

x′v

BA

Focal plane

(a) Direct

Focal plane

B

Aperturexlxs

xbxa Focal plane

Aperturexlxs

xbxa

x′a

T(xa,xb)

A BA

Out-of-focus path

(b) Indirect

Aperture

T(xa,xb)

xaxa B

AC

(c) In-focus indirect

Figure 4: We transform transient measurements H (xl,xs, t) at a visible relay wall (green) into virtual projectors (red) andcameras (blue) facing a hidden scene. We disentangle individual illumination components by computing elements of thevirtual transport matrix T(xa,xb): (a) We compute direct illumination (diagonal of T) by illuminating and imaging thesame location xv (xa ≡ xb); we gate light paths with the time of flight of 〈xl,xv,xs〉 to avoid defocusing effects fromempty voxels x′v . (b) Left: We compute indirect illumination (off-diagonal of T) by illuminating and imaging differentvoxels xa 6= xb, and gating illumination with the time of flight of 〈xl,xa,xb,xs〉. Right: Defocusing may still arise at theimaged voxel xb from paths with similar time of flight 〈xl,x

′a,xb,xs〉. (c) We isolate in-focus indirect light by using direct

illumination as an oracle of the hidden scene geometry (purple), and avoid imaging and illuminating empty locations.

therefore correspond to thin-lens propagators Lω(xl,xa)and Lω(xs,xb), respectively. This yields the following il-lumination function

Pω(〈xl,xa,xb,xs〉 , t) = G (ti,4, t)Lω(xl,xa)

|xl − xa|. (13)

Finally, we estimate the off-diagonal coefficientsToff,2(xa,xb) ≈ Ixa

(xb) as

Ixa(xb) =

∣∣∣∣∫S

∫L

[Pω(〈xl,xa,xb,xs〉 , t)∗H (xl,xs, t)]

Lω(xs,xb)

|xv − xs|dxldxs

∣∣∣∣2 . (14)

This represents the indirect contribution at xb from directillumination at xa after a single light bounce from xa to xb.To obtain the total indirect illumination of a single point xa

from the entire scene, we need to compute the correspond-ing column of Toff for all imaged points xb (yellow columnin Fig. 4b, left).

Discussion. While this method separates illuminationwith a time-of-flight ti,4, it is not necessarily restricted tothe path 〈xl,xa,xb,xs〉. As Fig. 4b (right) shows, whenfocusing illumination at an empty location xa, no light isreflected towards xb from xa. Instead, given the large illu-mination apertures L, light may be significantly defocusedover larger areas past xa (marked in red on object A). Thisin turn may lead to light paths 〈xl,x

′a,xb,xs〉with the same

time of flight ti,4 contributing to the imaged location xb

(large imaging apertures S may cause a similar problem).In the following we show how to mitigate these effects byleveraging our direct light estimations (Sec. 4.1) to isolatein-focus indirect illumination (i.e. surface to surface).

4.3. In-focus off-diagonal elements

While our gating procedure removes most defocused off-diagonal components, there may still be non-negligible illu-mination in Doff produced by focusing on empty regions ofthe scene. Direct-only illumination (Sec. 4.1) allows us toobtain an estimation Id (Eq. 12) of in-focus geometry in ourvoxelized space V . Since most empty regions have onlya very small contribution in Id(xv) (as a result of capturenoise), we can use Id(xv) as an oracle of the subspace Γ ofnon-empty voxels as

Γ = {xv ∈ V |Id(xv) > ε}, (15)

where ε represents a very small value.To constrain propagation in the transport matrix Toff,2

to in-focus indirect illumination, we use this subspace Γ toconstruct a 1D binary mask mΓ of size Kv, which we applyto columns and rows of Toff,2 (Fig. 4c). The resulting 2Dmask corresponds to the outer product

Mi = mΓ ⊗mΓ. (16)

We can therefore compute the indirect contribution of allvisible geometry using the same functions as in Sec. 4.2,limited to the subset of all non-empty locations xb ∈ Γ(Fig. 4c, purple), by traversing such non-empty locationsxa ∈ Γ and accumulating their contribution (Eq. 14) as

Ii (xb) =∑xa∈Γ

Ixa (xb) . (17)

This allows to accurately reconstruct in-focus two-bounce indirect illumination Toff,2 in a hidden scene evenin the presence of large virtual apertures, since we explicitlyavoid focusing light and camera on empty voxels.

2445

Page 7: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

Indirect illumination2nd bounce

No masking Masking

≥ 3rd bounce

No masking Masking

Figure 5: We illustrate effects of isolating indirect illumi-nation at surfaces by illuminating the center of a flat sur-face that reflects light towards a mannequin. Our maskingoperation using direct light as an oracle of geometry loca-tions removes most of the out-of-focus illumination in thescene, both in first-order (2nd-bounces) and higher-order (3or more bounces) indirect illumination.

Gating No gatingHidden scene

Figure 6: Our gating functions (center) are fundamentalto mitigate out-of-focus effects produced by large apertureconditions (right).

5. ResultsIn the following we demonstrate the performance of our

framework in both simulated and real scenes under differentscanning configurations of the impulse function H. All theresults shown here have been obtained using the illumina-tion and gating functions introduced in Sec. 4.1 to 4.3.

We compute the simulated scenes using a publicly avail-able transient renderer [14]. Unless stated otherwise, wecompute the impulse response H over a uniform 32 × 32-grid of SPAD positions and a 32×32-grid of laser positions;in both cases, they cover areas between 1×1 to 2×2 meterson the relay wall. This 2D-by-2D topology enables focus-ing both image and illumination, thus fully exploiting theprinciples of our framework.

Fig. 1 shows a complex hidden scene depicting acrowded living room. Probing the diagonal of T we com-pute the direct illumination component Td, which allowsus to recover the bodies at different depths (top), despiteocclusions and cluttering. We then isolate indirect lighttransport when illuminating different points in the scene(A, B and C) by probing their specific columns Toff (mid-dle). Last, masking out off-diagonal elements at differentdistances from the diagonal [32], we isolate indirect lighttransport within different path length intervals (bottom).

Fig. 5 shows how our masking procedure (Sec. 4.3) re-moves the contribution of defocus components Doff, for

32x32 SPAD array 32x1 SPAD array Single SPAD point

Hidden scene

Front projection

Volume

SPAD array topology32x32 SPAD 32x1 SPAD 1x1 SPAD

Figure 7: Examples with simpler capture setups consistingof 1D arrays and a single SPAD. The resulting LTMs leadto slightly degraded but consistent results.

Indirect (ours) Indirect (LOS)

Reflector specularity Reflector specularity

Figure 8: Comparison of our NLOS framework against areference LOS simulation observing the scene in Fig. 5, left,varying the specularity of the reflector. Our method is con-sistent with the LOS reference data.

2nd-bounce and higher-order indirect illumination. We il-luminate the center of a planar reflector to scatter indirectlight towards a mannequin. By using direct light as an ora-cle of geometry locations we remove most out-of-focus il-lumination in our indirect light estimations.

Our framework can also be used with simpler capturesetups, such as 1D arrays or even single SPADs (Fig. 7).In the case of 1D SPAD arrays the illumination focuses onstraight lines perpendicular to the orientation of the array.The resulting LTMs lead to slightly degraded but consistentseparation of illumination.

Validation: In Fig. 8 we evaluate qualitatively our off-diagonal estimation Toff, including the removal of the de-focus component Doff, by comparing against a LOS ren-der of the hidden scene (Fig. 5, left) for increasing reflectorspecularity. Our NLOS estimation (left) shows comparableresults with the LOS ground truth (right). Note that validat-ing virtual (phasor field) LOS against actual (optical) LOSsetups is challenging, since some scene features fall into thenull measurement space and thus cannot be recovered [19].

Real captured scenes: In our real captures our illumina-tion source consists of a OneFive Katana HP pulsed laseroperating at 532 nm and 700 mW, scanning a 24× 24 gridover a 1.9 × 1.9 meters area on the relay wall. We usea 1D SPAD array at the wall center, with 28 pixels hor-

2446

Page 8: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

Front Front TopTopDirect

Hidden sceneIndirect

Figure 9: Two real scenes, captured with a 2D laser grid anda 1D SPAD array. Blue marks indicate the coordinate wherelight is focused for indirect illumination computation. Ourframework allows to extract the direct component and iso-late the indirect transport from specific points in the scene,even from distant objects.

izontally spanning 15 cm. A PicoQuant HydraHarp 400Time-Correlated Single Photon Counting (TCSPC) recordsthe signal coming from the SPADs. The effective temporalresolution of our system is 85 ps. The total exposure timewas set to 50 seconds per scene. Fig. 9 shows the resultsfor two different scenes. Using a horizontal 1D SPAD ar-ray focuses illumination over a vertical line (dotted line inthe figure) instead of a point. Despite this operational lim-itation, our method is capable of sharply imaging both thedirect and indirect components of light transport in complexhidden scenes, including indirect transport from distant ob-jects (such as the circular panel in the second scene).

6. ConclusionsWe have introduced a framework to obtain the light

transport matrix of hidden scenes, coupling the well-established LOS transport matrix with recent NLOS for-ward propagation methods. This enables direct applicationof many existing LOS techniques in NLOS imaging setups.

We overcome the challenges posed by the wide aper-ture conditions of NLOS configurations by designing vir-tual imaging systems that mimic traditional narrow-apertureprojector-camera setups. We enable tailoring virtual illumi-nation functions to capture the diagonal and off-diagonalelements of the light transport matrix, roughly represent-ing direct and indirect illumination (first- and higher-orderbounces) in the hidden scene, respectively. In addition, wedemonstrate probing techniques for extracting different il-lumination components of hidden scenes featuring clutter-ing and lots of occlusions, as well as far/near-range decom-position of indirect light transport. Our framework can bedirectly used for improving NLOS reconstructions, or ana-lyzing the reflectance of hidden surfaces.

Limitations and future work: Our work shares resolu-tion limitations similar to other active illumination NLOSmethods. Like them, the topology and scanning density

of the captured impulse response define the maximum spa-tial frequency of our virtual projectors and cameras. Whilewe remove most out-of-focus residual from the off-diagonalcomponents, direct light may lead to overestimation of in-direct components due to ambiguous out-of-focus pathsfalling at geometry locations. Also, our current implemen-tation computes the LTM by brute-force sampling each col-umn, resulting in a cost of O(K2

v ). Although the com-putation of T can be restricted to diagonals and selectedcolumns, analyzing the properties of the virtual LTM (fol-lowing e.g., [41, 34, 30, 27]) and exploiting line focusingusing a 1D SPAD array (see Fig. 7) are interesting avenuesof future work for dramatically speeding up computations.

We have shown that our method works well with cur-rently available 1D SPAD arrays or even single-pixel sen-sors; the full deployment of 2D SPAD arrays (currentlyprototypes) will be instrumental in migrating classic LOSimaging systems to NLOS scenarios. Combining our for-mulation with the design of new scanning patterns [22] andbetter SPAD models [12] could lead to optimal light-sensortopologies to improve the estimation of transport matrices inhidden scenes. Last, designing new illumination functionsto isolate higher order bounces remains an exciting avenueof future work, which may extend the capabilities of NLOSsystems to enable imaging objects around a second corner.

AcknowledgementsWe would like to thank the reviewers for their feedback

on the manuscript, as well as Ibon Guillen for discussionsat early stages of the project. This work was funded byDARPA (REVEAL HR0011-16-C-0025), the European Re-search Council (ERC) under the European Union’s Horizon2020 research and innovation programme (CHAMELEONproject, grant agreement No 682080), the Ministry of Sci-ence and Innovation of Spain (project PID2019-105004GB-I00), and the University of Wisconsin-Madison Office of theVice Chancellor for Research and Graduate Education withfunding from the Wisconsin Alumni Research Foundation.

A. Gating higher-order bouncesHere we define a heuristic gating function to com-

pute additional off-diagonal elements correspondingto higher-order bounces Toff,3−∞. It selects paths〈xl,xa, . . . ,xb,xs〉 with a longer time-of-flight than the4-vertex paths ti,4 as

G (ti,3−∞, t) =

{0 t ≤ ti,41−G (ti,4, t) t > ti,4

, (18)

which is used in Eq. 13. Note that since the illumi-nated and imaged points xa and xb remain unchanged(i.e. columns and rows of Toff,3−∞), the propaga-tion uses the same thin-lens operators Lω(xl,xa) andLω(xs,xb).

2447

Page 9: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

References[1] Byeongjoo Ahn, Akshat Dave, Ashok Veeraraghavan, Ioan-

nis Gkioulekas, and Aswin C Sankaranarayanan. Convolu-tional approximations to the general non-line-of-sight imag-ing operator. In IEEE International Conference on ComputerVision (ICCV), pages 7889–7899, 2019. 2

[2] Victor Arellano, Diego Gutierrez, and Adrian Jarabo. Fastback-projection for non-line of sight reconstruction. Opt. Ex-press, 25(10), 2017. 2, 5

[3] Mauro Buttafava, Jessica Zeman, Alberto Tosi, Kevin Eli-ceiri, and Andreas Velten. Non-line-of-sight imaging usinga time-gated single photon avalanche diode. Opt. Express,23(16), 2015. 5

[4] Wenzheng Chen, Fangyin Wei, Kiriakos N. Kutulakos, Szy-mon Rusinkiewicz, and Felix Heide. Learned feature embed-dings for non-line-of-sight imaging and recognition. ACMTrans. Graph., 39(6), 2020. 2

[5] Paul Debevec, Tim Hawkins, Chris Tchou, Haarm-PieterDuiker, Westley Sarokin, and Mark Sagar. Acquiring thereflectance field of a human face. In SIGGRAPH, 2000. 1, 2

[6] Genevieve Gariepy, Nikola Krstajic, Robert Henderson,Chunyong Li, Robert R Thomson, Gerald S Buller, BarmakHeshmat, Ramesh Raskar, Jonathan Leach, and Daniele Fac-cio. Single-photon sensitive light-in-fight imaging. NatureCommunications, 6, 2015. 2

[7] Ibon Guillen, Xiaochun Liu, Andreas Velten, Diego Gutier-rez, and Adrian Jarabo. On the effect of reflectance onphasor field non-line-of-sight imaging. In IEEE Interna-tional Conference on Acoustics, Speech and Signal Process-ing (ICASSP), pages 9269–9273. IEEE, 2020. 3

[8] Mohit Gupta, Shree K Nayar, Matthias B Hullin, and JaimeMartin. Phasor imaging: A generalization of correlation-based time-of-flight imaging. ACM Trans. Graph. (ToG),34(5), 2015. 2

[9] Felix Heide, Matthias Hullin, James Gregson, and Wolf-gang Heidrich. Low-budget transient imaging using photonicmixer devices. ACM Trans. Graph., 32(4), 2013. 2

[10] Felix Heide, Matthew O’Toole, Kai Zang, David B Lindell,Steven Diamond, and Gordon Wetzstein. Non-line-of-sightimaging with partial occluders and surface normals. ACMTrans. Graph., 38(3), 2019. 1, 2

[11] Felix Heide, Lei Xiao, Andreas Kolb, Matthias B Hullin, andWolfgang Heidrich. Imaging in scattering media using cor-relation image sensors and sparse convolutional coding. Opt.Express, 22(21), 2014. 2

[12] Quercus Hernandez, Diego Gutierrez, and Adrian Jarabo.A computational model of a single-photon avalanchediode sensor for transient imaging. arXiv, preprintarXiv:1703.02635, 2017. 8

[13] Julian Iseringhausen and Matthias B Hullin. Non-line-of-sight reconstruction using efficient transient rendering. ACMTrans. Graph., 39(1), 2020. 2

[14] Adrian Jarabo, Julio Marco, Adolfo Munoz, Raul Buisan,Wojciech Jarosz, and Diego Gutierrez. A framework for tran-sient rendering. ACM Trans. Graph., 33(6), 2014. 7

[15] Adrian Jarabo, Belen Masia, Julio Marco, and Diego Gutier-rez. Recent advances in transient imaging: A computer

graphics and vision perspective. Visual Informatics, 1(1),2017. 2

[16] Ahmed Kirmani, Tyler Hutchison, James Davis, and RameshRaskar. Looking around the corner using transient imag-ing. In IEEE International Conference on Computer Vision(ICCV), 2009. 2

[17] Martin Laurenzis and Andreas Velten. Nonline-of-sight lasergated viewing of scattered photons. Opt. Eng., 53(2), 2014.2, 5

[18] David B Lindell, Gordon Wetzstein, and Matthew O’Toole.Wave-based non-line-of-sight imaging using fast fk migra-tion. ACM Trans. Graph., 38(4), 2019. 2

[19] Xiaochun Liu, Sebastian Bauer, and Andreas Velten. Anal-ysis of feature visibility in non-line-of-sight measurements.In The IEEE Conference on Computer Vision and PatternRecognition (CVPR), June 2019. 7

[20] Xiaochun Liu, Sebastian Bauer, and Andreas Velten. Pha-sor field diffraction based reconstruction for fast non-line-of-sight imaging systems. Nature Communications, 11(1),2020. 2, 3

[21] Xiaochun Liu, Ibon Guillen, Marco La Manna, Ji HyunNam, Syed Azer Reza, Toan Huu Le, Adrian Jarabo, DiegoGutierrez, and Andreas Velten. Non-line-of-sight imagingusing phasor fields virtual wave optics. Nature, 2019. 1, 2,3, 4, 5

[22] Xiaochun Liu and Andreas Velten. The role of Wigner dis-tribution function in non-line-of-sight imaging. In IEEEInternational Conference on Computational Photography(ICCP), 2020. 4, 8

[23] Tomohiro Maeda, Guy Satat, Tristan Swedish, LagnojitaSinha, and Ramesh Raskar. Recent advances in imagingaround corners. arXiv preprint arXiv:1910.05613, 2019. 1

[24] Yasuhiro Mukaigawa, Yasushi Yagi, and Ramesh Raskar.Analysis of light transport in scattering media. In IEEE Com-puter Vision and Pattern Recognition (CVPR), pages 153–160. IEEE, 2010. 2

[25] Shree K Nayar, Gurunandan Krishnan, Michael D Gross-berg, and Ramesh Raskar. Fast separation of direct andglobal components of a scene using high frequency illumi-nation. ACM Trans. Graph., 25(3), 2006. 1, 2, 3

[26] Ren Ng, Ravi Ramamoorthi, and Pat Hanrahan. All-frequency shadows using non-linear wavelet lighting approx-imation. ACM Trans. Graph., 22(3), 2003. 1, 3

[27] Matthew O’Toole, Supreeth Achar, Srinivasa G Narasimhan,and Kiriakos N Kutulakos. Homogeneous codes for energy-efficient illumination and imaging. ACM Trans. Graph.,34(4), 2015. 2, 3, 8

[28] M. O’Toole, F. Heide, D. Lindell, K. Zang, S. Diamond,and G. Wetzstein. Reconstructing transient images fromsingle-photon sensors. In IEEE Computer Vision and Pat-tern Recognition (CVPR), 2017. 2

[29] Matthew O’Toole, Felix Heide, Lei Xiao, Matthias B. Hullin,Wolfgang Heidrich, and Kiriakos N. Kutulakos. Temporalfrequency probing for 5D transient analysis of global lighttransport. ACM Trans. Graph., 33(4), 2014. 1, 2, 3

[30] Matthew O’Toole and Kiriakos N Kutulakos. Optical com-puting for fast light transport analysis. ACM Trans. Graph.,29(6), 2010. 1, 2, 3, 8

2448

Page 10: Virtual Light Transport Matrices for Non-Line-of-Sight Imaging

[31] Matthew O’Toole, David B Lindell, and Gordon Wetzstein.Confocal non-line-of-sight imaging based on the light-conetransform. Nature, 555(7696), 2018. 2

[32] Matthew O’Toole, Ramesh Raskar, and Kiriakos N Kutu-lakos. Primal-dual coding to probe light transport. ACMTrans. Graph., 31(4), 2012. 1, 2, 3, 4, 5, 7

[33] Adithya Pediredla, Akshat Dave, and Ashok Veeraraghavan.Snlos: Non-line-of-sight scanning through temporal focus-ing. In 2019 IEEE International Conference on Computa-tional Photography (ICCP), pages 1–13. IEEE, 2019. 2, 5

[34] Pieter Peers, Dhruv K Mahajan, Bruce Lamond, AbhijeetGhosh, Wojciech Matusik, Ravi Ramamoorthi, and Paul De-bevec. Compressive light transport sensing. ACM Trans.Graph., 28(1), 2009. 2, 3, 8

[35] Pieter Peers, Karl vom Berge, Wojciech Matusik, Ravi Ra-mamoorthi, Jason Lawrence, Szymon Rusinkiewicz, andPhilip Dutre. A compact factored representation of hetero-geneous subsurface scattering. ACM Trans. Graph. (TOG),25(3), 2006. 1

[36] Pradeep Sen, Billy Chen, Gaurav Garg, Stephen RMarschner, Mark Horowitz, Marc Levoy, and Hendrik PALensch. Dual photography. ACM Trans. Graph., 3(24), 2005.1, 2

[37] Pradeep Sen and Soheil Darabi. Compressive dual photog-raphy. Computer Graphics Forum, 28(2), 2009. 2

[38] Merrill Skolnik. Introduction to Radar Systems. 3rd Edition.McGraw-Hill Education, 2002. 4

[39] Andreas Velten, Thomas Willwacher, Otkrist Gupta, AshokVeeraraghavan, Moungi G. Bawendi, and Ramesh Raskar.Recovering three-dimensional shape around a corner usingultrafast time-of-flight imaging. Nature Communications,(3), 2012. 1, 2, 5

[40] Andreas Velten, Di Wu, Adrian Jarabo, Belen Masia,Christopher Barsi, Chinmaya Joshi, Everett Lawson, MoungiBawendi, Diego Gutierrez, and Ramesh Raskar. Femto-photography: Capturing and visualizing the propagation oflight. ACM Trans. Graph., 32(4), 2013. 2

[41] Jiaping Wang, Yue Dong, Xin Tong, Zhouchen Lin, andBaining Guo. Kernel Nystrom method for light transport.ACM Trans. Graph., 28(4), 2009. 1, 2, 3, 8

[42] Di Wu, Andreas Velten, Matthew O’Toole, Belen Masia,Amit Agrawal, Qionghai Dai, and Ramesh Raskar. Decom-posing global light transport using time of flight imaging.International Journal of Computer Vision, 107(2), 2014. 2

[43] Di Wu, Gordon Wetzstein, Christopher Barsi, ThomasWillwacher, Matthew O’Toole, Nikhil Naik, Qionghai Dai,Kyros Kutulakos, and Ramesh Raskar. Frequency analy-sis of transient light transport with applications in bare sen-sor imaging. In European Conference on Computer Vision(ECCV), 2012. 2

[44] Rihui Wu, Adrian Jarabo, Jinli Suo, Feng Dai, YongdongZhang, Qionghai Dai, and Diego Gutierrez. Adaptivepolarization-difference transient imaging for depth estima-tion in scattering media. Optics Letters, 43(6), 2018. 2

[45] Shumian Xin, Sotiris Nousias, Kiriakos N Kutulakos,Aswin C Sankaranarayanan, Srinivasa G Narasimhan, and

Ioannis Gkioulekas. A theory of Fermat paths for non-line-of-sight shape reconstruction. In IEEE Computer Vision andPattern Recognition (CVPR), 2019. 2

[46] Zexiang Xu, Kalyan Sunkavalli, Sunil Hadap, and RaviRamamoorthi. Deep image-based relighting from optimalsparse samples. ACM Trans. Graph., 37(4), 2018. 2

[47] Sean I Young, David B Lindell, Bernd Girod, David Taub-man, and Gordon Wetzstein. Non-line-of-sight surface re-construction using the directional light-cone transform. InIEEE Conference on Computer Vision and Patter Recogni-tion (CVPR), 2020. 2

2449