Deep Parametric Indoor Lighting Estimation

10/19/2019 ∙ by Marc-André Gardner, et al. ∙ 13

We present a method to estimate lighting from a single image of an indoor scene. Previous work has used an environment map representation that does not account for the localized nature of indoor lighting. Instead, we represent lighting as a set of discrete 3D lights with geometric and photometric parameters. We train a deep neural network to regress these parameters from a single image, on a dataset of environment maps annotated with depth. We propose a differentiable layer to convert these parameters to an environment map to compute our loss; this bypasses the challenge of establishing correspondences between estimated and ground truth lights. We demonstrate, via quantitative and qualitative evaluations, that our representation and training scheme lead to more accurate results compared to previous work, while allowing for more realistic 3D object compositing with spatially-varying lighting.



There are no comments yet.


page 1

page 3

page 5

page 6

page 7

page 8

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Recovering the lighting in a scene from a single image is a highly ill-posed problem. Since images are formed by conflating lighting with surface reflectance, scene geometry and the camera response function, inverting the image formation process to recover any of these components is severely under-constrained. This is especially true when the image has low dynamic range (LDR) and limited field of view, such as one captured by standard consumer cameras.

In his pioneering work in image-based lighting, Debevec [6] proposed to directly capture the lighting conditions at a location in the image by inserting a light probe at that location. The resulting HDR environment map represents the illumination incident from every direction at that point in the scene

, and can be used to realistically relight virtual objects at that location. Recently, fully automatic methods leveraging deep learning were proposed to estimate an environment map from a single indoor image 

[7]. However, the environment map representation assumes that lighting is distant — which is why it can be represented as a function of the incident direction. This assumption is often violated for indoor scenes that have localized light sources leading to spatially-varying lighting in the scene. Consequently, using a single environment map for the entire scene leads to inconsistent results for applications such as 3D object compositing (see Figure 1(a)).

            (a) Gardner et al. [7]                (b) Ours
Figure 1: Our method takes an indoor image as input and estimates a parametric lighting representation, enabling tasks such as virtual object insertion. Unlike previous methods that predict global lighting [7] (a), our estimates vary based on spatial location (b), generating different shadow directions (green, teal) and shading (orange, blue) that realistically adapt to the location in the image.

Alternatively, parametric lighting models represent illumination using a discrete set of light sources with various geometric and photometric properties. This is a global representation that can be used to reconstruct lighting at any 3D location in the scene (assuming known scene geometry). Techniques for learning parametric lighting have been proposed for the simpler case of outdoor illumination [13, 10]. Indoor lighting, on the other hand, is much more complex, with a varying number of light sources, all with potentially different properties, and located in close proximity to the scene. As a result, techniques for recovering parametric lighting in indoor scenes rely on extensive user input [11]

or hand-crafted heuristics that can often fail 

[12, 3].

In this paper, we propose a learning-based method for estimating parametric lighting from a single indoor image. In particular, our method produces a set of lighting parameters describing area lights distributed in 3D, from a single 2D LDR image. Specifically, our method predicts the light source positions (in 3D), areas, intensities and colors. These parameters can be directly plugged into a rendering engine to relight an object at any location in the scene, leading to photorealistic results as in Figure 1(b). Moreover, because these parameters are directly related to the physical properties of lighting, they provide an intutitive space for artists to manipulate and design the lighting they want for the scene.

Our network is trained on a large dataset of indoor HDR environment maps [7]

that we have manually labelled with pixel-wise depth information and annotated with our parametric lights. We encode an input RGB image into a latent feature vector from which we decode the parameters for a fixed number of light sources. Compared to Gardner et al. 

[7] who regress a full environment map, this is a significantly smaller set of parameters, leading to better performance while training on a much smaller dataset.

One approach to train our network would be to penalize the differences between the predicted lights and the annotated lights. However, this requires establishing a correspondence between these two sets of lights. This is fragile, especially earlier in training when the estimated light sources are arbitrary, and is unstable to train. Instead, we utilize a differentiable (non-learnable) layer that converts the predicted light parameters to an environment map. This allows us to compute the loss directly with respect to the environment maps without requiring any correspondences between light sources. Once trained in this manner, the network predicts good positional information, which we use to establish correspondences and further fine-tune the parameters. We demonstrate that this yields illumination estimates that are superior to previous lighting estimation methods.

Contributions. In summary, our main contributions are:

  1. [nosep,leftmargin=*]

  2. A deep neural network to estimate parametric 3D lighting from a single indoor image;

  3. A dataset of depth annotations over panoramas that can be used to learn 3D scene lighting;

  4. A robust training procedure based on a differentiable parametric loss with respect to environment maps.

2 Related Work

Illumination estimation is a classic vision/graphics problem and is a critical component of work on scene reconstruction and rendering. A large body of work on illumination estimation focuses on reconstructing lighting from images of individual objects. Some of these methods assume that the shape of the object is either known [21, 18] or can be reconstructed using shape priors [4] or heuristics [19]. Barron and Malik [2] recover geometry, reflectance and illumination from a single image of an arbitrary object by enforcing hand-crafted priors on each component. Recently, deep learning-based methods have been proposed to recover illumination and material properties (along with, in some cases, geometry) from a single RGB image of an object [8, 17, 22, 16]. These methods do not easily scale to large-scale indoor scenes where the illumination, geometry, and reflectance properties are significantly more complex.

Methods for estimating lighting in large-scale indoor scenes often assume known geometry. Barron and Malik [3] and Maier et al. [20] assume an RGBD input image and recover spatially-varying Spherical Harmonics illumination. Zhang et al. [26] reconstruct parametric 3D lighting but require a full multi-view 3D reconstruction of the scene. Karsch et al. [11] recover parametric 3D lighting from a single image, but do so by requiring substantial user input to reconstruct coarse geometry and initialize the lighting.

Lalonde et al. [13] uses hand-crafted priors based on scene cues such as shadows to recover lighting from a single outdoor image. Hold-Geoffroy et al. [10] propose a deep neural network that does the same. Both of these methods utilize low-dimensional analytical outdoor illumination models; indoor illumination is significantly more complex.

Similar to us, Karsch et al. [12] estimate parametric 3D lighting from a single indoor image. They estimate scene geometry and reflectance, detect light source positions, and then estimate light source intensities using a rendering-based optimization. Each subtask in their pipeline is a challenging inverse problem that is solved using heuristics; errors in each component can propagate forward leading to inaccurate lighting estimates. In contrast, Gardner et al. [7] propose an end-to-end deep network to regress lighting, represented as an environment map, from a single image. They train this network on a large-scale dataset of LDR environment maps [25] and fine-tune it on an HDR environment map dataset. Legendre extend this work to mobile applications and obtain better results by using a collection of videos as training data [14]. Song   [23] use Matterport3D [5] dataset and a novel warping procedure in order to support multiple insertion points. We improve on their work by training an end-to-end neural network to predict discrete parametric 3D lights with 3D position, area, color and intensity. We show that this lighting model comes with many advantages. First, it is a compact representation that is easier to learn; we show that training only on the HDR environment map dataset leads to more accurate lighting predictions than their work. Second, a 3D representation naturally handles spatially-varying lighting at different scene points. Finally, we can give artists access to individual light source parameters, allowing them to intuitively edit scene lighting.

3 Method

We aim to predict lighting conditions from a single, low dynamic range image. We frame this as the following learning problem: given input image , predict a set of parameters which accurately represents the illumination conditions in . Let be a set of lights and an ambient term:


where is the ambient term in RGB. Each light is represented by four parameters:


where is a unit vector specifying the direction of the light in XYZ coordinates (we found that encoding direction as a 3-vector was more stable to train compared to spherical coordinates that have problems with wrap-around), is a scalar encoding the distance in meters, the angular size of the light in steradians, and the light source color in RGB. Here, , and are defined with respect to the camera. While may vary from one image to another, in practice we set to a fixed value (e.g. ) since it is always possible to effectively “remove” a light by setting . We demonstrate this in our results in fig. 4.

Figure 2: Comparison between an environment map and our lighting representation for relighting. Top row: source HDR panorama, and spherical gaussian representation of the parameters. Second row: renders generated by both representations. Third and fourth row: lighting and renders when moving in the scene, using EnvyDepth [1] as ground truth. Note how our parametric lights stay true to the ground truth for a different location in the scene.

3.1 A Parametric Indoor Lighting Dataset

To train a deep lighting estimation network, we would ideally need a large dataset of images with labelled ground truth 3D light sources. Unfortunately, no such dataset exists. While the Matterport 3D dataset [5] has a diverse set of indoor scenes with HDR panoramas, the tops of these panoramas (up to from the zenith) are missing; given that a significant portion of indoor illumination comes from the ceilings, this is an issue for lighting estimation methods. Another option would be to use synthetic indoor scene datasets such as SUNCG [24]; however, SUNCG has unrealistic material maps that do not match real world appearance. Instead, we rely entirely on real data for training. For this, we use the Laval Indoor HDR Dataset [7]111Available at, which contains 2,100 HDR panoramas taken in a variety of indoor environments. We manually annotated each panorama in this dataset using EnvyDepth [1] to obtain per-pixel depth estimates. As with [7], we extract eight limited field-of-view crops for each panorama to form our dataset.

We retrieve the ground truth light sources in these panoramas with a simple algorithm. Here, there is no need to train light source detectors as in [7, 12] since the dataset is HDR. We first extract the peak value of the panorama. Then, simple region detection is employed to initialize seeds, which are then grown until the intensity goes under a third of the peak. Other lights are detected by repeating the process after masking the detected sources, until an energy threshold is met. The set of light parameters is obtained by computing the corresponding on each light source independently: is the vector from the camera position pointing towards the center of mass of the light pixels, is the mean depth obtained from EnvyDepth, is the average angular size of the major and minor axes of an ellipse fitted on the light and is the mean RGB color of the light pixels. Fig. 2 shows the result of applying this algorithm on real HDR panoramas from the Laval Indoor HDR Dataset, and illustrates that doing so results in relighting results that are close to those obtained with the input environment map.

We then process the dataset to retrieve for all images. At this step, varies depending on the content of each environment map, ensuring that models every significant light source (a light is deemed as significant if it is providing at least 10% of the energy of the strongest source). Once extracted, light intensities are fine-tuned using a rendering-based optimization process. We render a simple object using the environment map masked everywhere except for the detected lights. We then render the same object using each of our extracted parametric lights. We optimize the parametric light colors and intensities such that they minimize the error with respect to the ground truth environment map.

3.2 Network architecture

Figure 3: Proposed network architecture. A blue box indicates an output layer. Output sizes vary depending on , except for .

We learn the mapping from input image to a set of estimated lighting parameters using a deep neural network, whose architecture is shown in fig. 3. Specifically, the input image goes through a headless DenseNet-121 network to produce a 3072-dimensional latent vector . The vector is then forwarded to another fully-connected layer (512 units), common to all parameter decoders. From this layer, output layers are defined: four outputting the light source parameters, and one for the ambient term . Since they are trained separately (sec. 3.3), the layers also take in the position predictions as input.

This network architecture has two key differences compared to the state-of-the-art, non-parametric lighting estimation method of Gardner et al. [7]. First, while [7] required decoders to produce HDR intensity and RGB from the latent vector , we instead use simple parametric decoders to directly infer from

. Doing so offers the advantage of reducing the number of parameters needed in the decoder, and results in a much faster network. The second key difference is that we leverage a standard feature extractor, DenseNet121 pre-trained on ImageNet, as the encoder. We found this made for more stable learning, avoids a computationally costly pretraining step, and enables us to train solely on the Laval Indoor HDR Dataset (as opposed to

[7] which also had to train on SUN360 LDR panoramas [25]).

(a) Input (b) GT Environment map (c) 2 lights (d) 3 lights (e) 5 lights
Figure 4: Lighting predictions for networks trained with different numbers of lights . From left to right: input image, ground truth lighting, and predictions for networks trained with respectively. When trained to output more lights, the network is able to effectively turn off some lights by decreasing their intensity in order to fit a scene requiring a lesser number of light sources.

3.3 Training procedure

While using a parametric modeling of lighting had the advantage of simplifying the neural network architecture, it imposes significant complications on the training process. Indeed, how can we compute a loss on a discrete set of light sources when many of these light sources are not even visible in the image?

For this, one could dynamically assign each predicted light to the closest ground truth light source according, for example, to their angular distance. However, this creates a dependency between the estimated light position and all the other parameters since the assignment is made based on . This causes problems earlier in the training process, when predicted light directions could change arbitrarily. This creates “poor” light assignments that, in turn, lead to unstable gradient flow and network convergence. It also creates ambiguities, such as when two estimated light sources are close to a single ground truth light, etc. We find that training in such a way results in RMSE and si-RMSE several times over the results reported in sec. 4, that use the following two-step training procedure.

3.3.1 Training step 1: radius, color and position

In the first training step, we bypass the need for assigning predicted lights to ground truth light sources by rendering an environment map from the parameters and comparing it to a ground truth environment map:


where is the ground truth environment map associated with , is a function mapping to a environment map, and is the L2 loss. To obtain the ground truth environment map , we use the same warping operator as in [7], and threshold the warped environment map to 5% of its peak intensity value, in order to keep only the brightest light sources. The remaining energy is averaged and assigned to the ground truth “ambient term”, . Since we care more about detecting the important light sources in the scene, we scale the losses for the light sources and the ambient term differently; we set and .

The differentiable projection function could be implemented via existing differentiable renderers [15], however we experimentally found them to be unstable when the parameters are far from their target, in addition to requiring significant computations for each scene. Instead, we implement by projecting each light source on the sphere onto a spherical gaussian using the mapping


where is a unit vector giving the direction on the sphere. The angular size is scaled such that the light intensity falls under 10% of its peak after this distance. can be computed efficiently since can be precomputed for all possible values of . Moreover,

is differentiable everywhere on the sphere, allowing us to backpropagate the error in the environment reconstruction to the predicted 3D light parameters and to the rest of the network.

Our parametric lighting-to-environment map projection function plays an important role in our method. As noted above, parametric lights give us a compact, low-dimensional representation of scene illumination as well as access to individual lights in the scene. However, an assignment-based optimization approach that relies on correspondences between predicted and ground truth light sources is akin to a greedy, “local” optimization scheme which might have bad convergence behaviour. Converting our parametric lights to a environment map allows us to minimize error “globally” and leads to significantly better convergence. It also implicitly handles issues such as a mismatch between the number of predicted lights and ground truth lights, thereby making it possible for us to train our network with any number of preset light sources.

Fig. 2 shows an example of an environment map and its equivalent parametric representation using our projection function. The figure also shows how this parametric representation can be projected at different locations in the scene and account for the variations in lighting.

3.3.2 Training step 2: depth and refinement

Once the network has converged using the first training step, we find that the network predicts good light source positions, , and we switch to independent parametric losses since the assignment problem raised above no longer applies. In this stage we only fine-tune the individual parametric heads and keep the network frozen up to the 512-d feature vector. The position head is also frozen to avoid any unwanted feedback. We add a depth estimation head and condition it on both the DenseNet features,

, and the light source position. We use the following loss function:


where the L2 losses above are computed with respect to the closest light source (angular distance) in the ground truth set . If a light source has an angular distance greater than , then no loss is computed as it is deemed too far away from any ground truth light source to be reliable.

3.4 Implementation details

As mentioned, we use a pretrained DenseNet-121 as encoder for our network. Our entire network contains 8.5M parameters, or about fewer than [7]. During the first training step (sec. 3.3.1

), the network is trained for 150 epochs, with an early stopping mechanism based on a small validation set. In the second step (sec. 

3.3.2), the network is trained for an additional 50 epochs. In both cases, the Adam optimizer is used with a learning rate of (halved each 25 epochs) and . The first training pass typically takes 12 hours before convergence on a Titan V GPU, while the second finetuning pass completes in a few hours. A batch size of 48 samples was used.

One advantage of our proposed method over previous work is its efficiency in terms of compute requirements. Our parametric representation requires simpler decoders than the ones required to generate a full environment map as in [7]. As such, our method is much faster than previous work at test time, executing in roughly 51ms per image on a CPU, compared to 127ms for [7] or 5 minutes for [12]. Our method can also be parallelized on GPU, processing 48 images in around 16ms on an nVidia Titan V.

4 Evaluation

1st percentile

25th percentile

50th percentile

75th percentile

(a) Input image (b) GT EnvyDepth panorama (c) Our predicted panorama (d) GT render (e) Our render
Figure 5: Examples of errors made by our network, at different RMSE percentiles. The spiky spheres shown on the right are lit by the lighting corresponding to the input image center. Note that even at large error percentiles (bottom), the corresponding renders are plausible.
Figure 6: Quantitative comparison between our method configured with 2, 3 and 5 lights and the Gardner et al. method [7] against the ground truth, as provided by EnvyDepth [1] for (left) RMSE and (right) scale-invariant (si-)RMSE. Results are shown for three insertion position: center and 1m to the left and right. Our method with lights is significantly better in both RMSE and SI-RMSE metrics, compared to [7], as asserted by Kruskal-Wallis, Mann-Whitney and Wilcoxon tests with .

Light estimation methods can be tricky to evaluate. In particular, quantitative evaluation of realism is not straightforward to define, since realism is in itself ambiguous and mainly relies on human perception. In this section, we present an extensive evaluation protocol which includes both quantitative and qualitative comparisons to the state-of-the-art indoor lighting estimation methods of Gardner et al. [7] and Karsch et al. [12]. We first compute the RMSE between renders made with lighting predictions against a ground truth surrogate (using Envydepth [1]). We then present the results of a user study to quantitatively compare the realism of our results against previous state-of-the-art.

On the qualitative side, we provide many examples of object insertion (including all 19 user study scenes) on a test set of 252 panoramas from the Laval Indoor HDR Dataset (selected so as to not overlap the training set). We also show virtual object insertion results on stock images (not from our dataset), to demonstrate our approach robustness. Finally, we provide 50 images containing two bunnies, one real and one relit using our prediction. This allows for a direct comparison with an object with a known appearance.

Additionally, our parametric output allows for intuitive user editing, as we demonstrate in supplementary material.

Ground truth Ours Karsch et al. [12] Gardner et al. [7]
Figure 7: Three examples taken from the user study (fig. 8). From left to right: the ground truth illumination (using ground truth HDR panorama and Envydepth [1]), our proposed method, Karsch et al. [12] and Gardner et al. [7]. Note how our method captures the light color and overall exposure well compared to other methods. All images used in the user study are provided in supplementary material.
Figure 8: User study results (49 participants, 19 scenes). The percentages denote the fraction of time each method was preferred to the ground truth illumination (perfect confusion = 50%).
Figure 9: Single image lighting estimation results comparing between parametric and non-parametric lighting. We compare environment maps and renders between a) the ground truth illumination, b) the non-parametric approach of Gardner et al. [7], and c) our parametric lighting representation. More results in supplementary material.

4.1 Quantitative evaluation

We quantitatively compare three different configurations of our method to Gardner et al. [7] on our test set. To do so, we render a diffuse virtual object—the spiky sphere shown on the right of fig. 5 that captures both shading and shadowing—at three different locations in the scene: in the center, 1 meter to the left, and 1 meter to the right. We do this for each method and the ground truth environment map, warped using the EnvyDepth geometry.

We evaluate the accuracy of each method by computing the RMSE and scale-invariant (si-RMSE) [9] metrics on the corresponding renderings; these are shown in fig. 6. RMSE is most sensitive to errors in overall light intensity. We observe that our technique significantly outperforms [7]

in this regard. Furthermore, we see a large improvement in performance as we increase the amount of lights estimated from 2 to 3, and finally 5. This result suggests that adding more lights gives more degrees of freedom to the network to better fit the overall energy distribution. Another improvement of our approach is that our parametrization allows for each light source to have a different color, while

[7] only predicts the intensity. Finally, our errors are consistent across different insertion points indicating our ability to adapt to scene location, unlike [7]

which exhibits higher variance.

The si-RMSE measure factors out scale differences between the different methods and ground truth, and focuses on cues such as shading and shadows, mainly due to light position. Our results show that performance increases from a 2 lights to a 3 lights configuration. However, performance slightly decreases from 3 to 5 lights, suggesting that optimizing the positions of 5 lights is harder than 3 lights. Our method with the 3-lights configuration obtains a 40% increase in median estimation performance on si-RMSE over [7]. Finally, we note that a network trained using a direct loss from the start (e.g. skipping the first training step, sec. 3.3.1) obtains si-RMSE errors an order of magnitude higher than with our two steps approach, effectively validating the light assignation issues reported in sec. 3.3. Fig. 4 shows a qualitative visualization of the environment maps recovered with a varying number of lights.

We also compare the methods’ realism using a user study. Users were shown pairs of images with composited objects—rendered either with ground truth lighting or a randomly selected method’s prediction—and asked to pick the more realistic image. We used 19 scenes with multiple virtual objects scattered in them. Results (49 participants) are shown in fig. 8 and confirm the improvement over [7] and [12]. Examples of these scenes are shown in fig. 7.

Figure 10: Relighting results comparing virtual and real bunnies. Images were captured using a smartphone (Pixel 2, HDR off) in various environments, including outdoors. Our approach can closely match the true lighting, even in situations it was not trained for like outdoor daylight, and reliably produces realistic results.
Figure 11: Object relighting on a variety of generic stock photos downloaded from the Internet. More results on such out-of-dataset images are provided in supplementary material.

4.2 Qualitative evaluation

Fig. 9 shows two representative qualitative comparisons to the non-parametric approach of Gardner et al. [7]. For each example, the estimated environment map and a render of diffuse and glossy objects are shown. To highlight the difference between parametric (ours) and non-parametric ([7]) lighting representations, the objects are lit by the lighting estimated at the center of the scene. Even though our representation is much more compact than that of [7], the rendering results are either qualitatively similar (first column) or better match ground truth (last column) than [7].

We also compare rendered and real versions of the same object in fig. 10. For each image, a real bunny model was placed in the scene. The images were acquired using a regular smartphone camera in order to further validate the robustness of the network with out-of-dataset samples. From the network estimation, we relight a virtual model of the same bunny and insert it close to the real one. Although the real and rendered bunnies slightly differ in appearance, the overall comparison demonstrates our method’s ability to recover correct light positions and intensity (more results on similar scenes in supplementary material).

Finally, in fig. 11, we demonstrate our method’s ability to generalize to stock photos, including some that have undergone artistic processing, and still produce realistic results.

5 Conclusion

We have presented a method to estimate parametric 3D lighting—comprised of discrete light sources with positions, area, color and intensity—from a single indoor image. Our lighting model allows us to render incident illumination at any location in a scene; this is critical for indoor scenes which often have localized lighting that cannot be accurately modeled by global lighting estimates. We train this method end-to-end using a differentiable parametric loss based on an environment map representation. We have demonstrated that our method is robust and significantly outperforms previous work in terms of lighting accuracy and allows for realistic virtual object insertion where objects are lit differently based on their insertion point.

Our work offers a number of directions for future exploration. First, while our lighting representation is 3D, using it requires a 3D scene reconstruction to specify locations where to estimate lighting and to compute light visibility. Second, our lighting model assumes diffuse area light-like sources and cannot model directional lights or focused light beams. Extending this model to handle more general sources would generalize it to more indoor scenes. Finally, our method focuses solely on estimating scene illumination. However, lighting is only one of many scene properties that affect its appearance. We hypothesize that jointly reasoning about scene appearance—including geometry, materials, and illumination—could improve overall accuracy.


We acknowledge the financial support of NSERC for the main author PhD scholarship. This work was supported by the REPARTI Strategic Network, the NSERC Discovery Grant RGPIN-2014-05314, MITACS, Prompt-Québec and E Machine Learning. We gratefully acknowledge the support of Nvidia with the donation of the GPUs used for this work, as well as Adobe with generous gift funding.


  • [1] F. Banterle, M. Callieri, M. Dellepiane, M. Corsini, F. Pellacini, and R. Scopigno (2013) EnvyDepth: an interface for recovering local natural illumination from environment maps. Computer Graphics Forum 32 (7), pp. 411–420. Cited by: Figure 2, §3.1, Figure 6, Figure 7, §4.
  • [2] J. Barron and J. Malik (2013) Shape, illumination, and reflectance from shading. IEEE Transactions on Pattern Analysis and Machine Intelligence 37 (8), pp. 1670–1687. Cited by: §2.
  • [3] J. Barron and J. Malik (2015) Intrinsic scene properties from a single rgb-d image. TPAMI. Cited by: §1, §2.
  • [4] V. Blanz and T. Vetter (1999) A morphable model for the synthesis of 3d faces. In SIGGRAPH, pp. 187–194. External Links: ISBN 0-201-48560-5 Cited by: §2.
  • [5] A. Chang, A. Dai, T. Funkhouser, M. Halber, M. Niessner, M. Savva, S. Song, A. Zeng, and Y. Zhang (2017) Matterport3D: learning from RGB-D data in indoor environments. International Conference on 3D Vision (3DV). Cited by: §2, §3.1.
  • [6] P. Debevec (1998) Rendering synthetic objects into real scenes: bridging traditional and image-based graphics with global illumination and high dynamic range photography. In Proceedings of the 25th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH, pp. 189–198. Cited by: §1.
  • [7] M. Gardner, K. Sunkavalli, E. Yumer, X. Shen, E. Gambaretto, C. Gagné, and J. Lalonde (2017) Learning to predict indoor illumination from a single image. ACM Transactions on Graphics (SIGGRAPH Asia) 9 (4). Cited by: Figure 1, §1, §1, §2, §3.1, §3.1, §3.2, §3.3.1, §3.4, §3.4, Figure 6, Figure 7, Figure 9, §4.1, §4.1, §4.1, §4.1, §4.2, §4.
  • [8] S. Georgoulis, K. Rematas, T. Ritschel, M. Fritz, T. Tuytelaars, and L. V. Gool (2017) What is around the camera?. In ICCV, Cited by: §2.
  • [9] R. Grosse, M. K. Johnson, E. H. Adelson, and W. T. Freeman (2009) Ground truth dataset and baseline evaluations for intrinsic image algorithms. In

    IEEE International Conference on Computer Vision

    Cited by: §4.1.
  • [10] Y. Hold-Geoffroy, K. Sunkavalli, S. Hadap, E. Gambaretto, and J. Lalonde (2017) Deep outdoor illumination estimation. In

    IEEE Conference on Computer Vision and Pattern Recognition

    Cited by: §1, §2.
  • [11] K. Karsch, V. Hedau, D. Forsyth, and D. Hoiem (2011) Rendering synthetic objects into legacy photographs. ACM Transactions on Graphics 30 (6), pp. 1. Cited by: §1, §2.
  • [12] K. Karsch, K. Sunkavalli, S. Hadap, N. Carr, H. Jin, R. Fonte, M. Sittig, and D. Forsyth (2014) Automatic scene inference for 3d object compositing. ACM Transactions on Graphics (3), pp. 32:1–32:15. Cited by: §1, §2, §3.1, §3.4, Figure 7, §4.1, §4.
  • [13] J. F. Lalonde, A. A. Efros, and S. G. Narasimhan (2012) Estimating the natural illumination conditions from a single outdoor image. International Journal of Computer Vision 98 (2), pp. 123–145. Cited by: §1, §2.
  • [14] C. LeGendre, W. Ma, G. Fyffe, J. Flynn, L. Charbonnel, J. Busch, and P. Debevec (2019-06) DeepLight: learning illumination for unconstrained mobile mixed reality. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Cited by: §2.
  • [15] T. Li, M. Aittala, F. Durand, and J. Lehtinen (2018) Differentiable monte carlo ray tracing through edge sampling. ACM Trans. Graph. (Proc. SIGGRAPH Asia) 37 (6), pp. 222:1–222:11. Cited by: §3.3.1.
  • [16] Z. Li, Z. Xu, R. Ramamoorthi, K. Sunkavalli, and M. Chandraker (2018-12) Learning to reconstruct shape and spatially-varying reflectance from a single image. ACM Trans. Graph. 37 (6), pp. 269:1–269:11. Cited by: §2.
  • [17] G. Liu, D. Ceylan, E. Yumer, J. Yang, and J. Lien (2017) Material editing using a physically based rendering network. Cited by: §2.
  • [18] S. Lombardi and K. Nishino (2016) Reflectance and illumination recovery in the wild. IEEE Transactions on Pattern Analysis and Machine Intelligence 38 (1), pp. 129–141. Cited by: §2.
  • [19] J. Lopez-Moreno, E. Garces, S. Hadap, E. Reinhard, and D. Gutierrez (2013) Multiple light source estimation in a single image. Computer Graphics Forum. Cited by: §2.
  • [20] R. Maier, K. Kim, D. Cremers, J. Kautz, and M. Nießner (2017-10) Intrinsic3D: high-quality 3D reconstruction by joint appearance and geometry optimization with spatially-varying lighting. In International Conference on Computer Vision (ICCV), Venice, Italy. Cited by: §2.
  • [21] S. R. Marschner and D. P. Greenberg (1997) Inverse lighting for photography. In IS&T Color Imaging Conference, Cited by: §2.
  • [22] A. Meka, M. Maximov, M. Zollhoefer, A. Chatterjee, H. Seidel, C. Richardt, and C. Theobalt (2018) LIME: live intrinsic material estimation. In IEEE Conference on Computer Vision and Pattern Recognition, Cited by: §2.
  • [23] S. Song and T. Funkhouser (2019-06) Neural illumination: lighting prediction for indoor environments. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Cited by: §2.
  • [24] S. Song, F. Yu, A. Zeng, A. X. Chang, M. Savva, and T. Funkhouser (2017) Semantic scene completion from a single depth image. IEEE Conference on Computer Vision and Pattern Recognition. Cited by: §3.1.
  • [25] J. Xiao, K. A. Ehinger, A. Oliva, and A. Torralba (2012) Recognizing scene viewpoint using panoramic place representation. In IEEE Conference on Computer Vision and Pattern Recognition, Cited by: §2, §3.2.
  • [26] E. Zhang, M. F. Cohen, and B. Curless (2016) Emptying, refurnishing, and relighting indoor spaces. ACM Transactions on Graphics 35 (6). Cited by: §2.