This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Acquisition of Spatially-Varying Reflectance and Surface Normals via Polarized Reflectance Fields

Jing Yang1, Pratusha Bhuvana Prasad1, Qing Zhang2, Yajie Zhao1
1University of Southern California, 2Sony R&D Center US
{jyang, bprasad, zhao}@ict.usc.edu
Abstract

Accurately measuring the geometry and spatially-varying reflectance of real-world objects is a complex task due to their intricate shapes formed by concave features, hollow engravings and diverse surfaces, resulting in inter-reflection and occlusion when photographed. Moreover, issues like lens flare and overexposure can arise from interference from secondary reflections and limitations of hardware even in professional studios. In this paper, we propose a novel approach using polarized reflectance field capture and a comprehensive statistical analysis algorithm to obtain highly accurate surface normals (within 0.1mm/px) and spatially-varying reflectance data, including albedo, specular separation, roughness, and anisotropy parameters for realistic rendering and analysis. Our algorithm removes image artifacts via analytical modeling and further employs both an initial step and an optimization step computed on the whole image collection to further enhance the precision of per-pixel surface reflectance and normal measurement. We showcase the captured shapes and reflectance of diverse objects with a wide material range, spanning from highly diffuse to highly glossy — a challenge unaddressed by prior techniques. Our approach enhances downstream applications by offering precise measurements for realistic rendering and provides a valuable training dataset for emerging research in inverse rendering. We will release the polarized reflectance fields of several captured objects with this work.

1 Introduction

Creating realistic renderings of real-world objects is a complex task with diverse applications, including online shopping, game design, VR/AR telepresence, and visual effects. It requires precise modeling and measurement of an object’s 3D geometry and reflectance properties. Recent advancements in neural renderings, such as NeRF [37] and Gaussian Splatting [28], offer superior realism through implicit representation but are limited to fixed scenes with fixed illumination. Ongoing research [42, 5, 6, 47] explores relighting and inverse-rendering in neural fields. This research requires an understanding of real-world object materials, necessitating a database of material measurements. However, accurately estimating 3D geometry and reflectance properties, encompassing diffuse and specular aspects, poses a significant challenge due to the complex interplay of lighting, geometry, and spatially varying reflectance [1]. Pioneer techniques [22, 46, 2] combine multi-view 3D reconstruction and photography under diverse illumination to measure the geometry and spatially-varying Bidirectional Reflectance Distribution Function (SVBRDF) of real-world objects. Once measured, the models become renderable from any perspective, enabling a faithful representation of digital models in virtual environments.

Acquiring an object’s reflectance properties involves measuring its SVBRDF, which requires observations from continuously changing view and lighting angles, resulting in large amounts of captured texture data. Previous works [11, 40, 45, 15] fall into this category, capturing 4D texture databases as Bidirectional Texture Functions (BTFs) and employing data-driven image-based re-rendering. These methods interpolate between sampled BRDF values from the captured 4D textures, offering accuracy but requiring substantial storage and custom interpolation functions. Moreover, they often focus on small patches, neglecting surface geometry [40], resulting in incompatibility with modern rendering engines. A recent discrete and sparse pattern-based approach [18] adopts a parametric SVBRDF, achieving similar rendering quality for fabric materials compared to BTFs with a more intuitive representation. However, they optimize reflectance via SSIM loss, which may suffer from high bias in local minima.

Most material reflectance capture methods employ analytical BRDF models that rely on a few parameters, enabling sparse observations for parameter estimation and seamless integration into modern rendering pipelines for realism. Pioneering practical techniques by [35, 21] utilize programmable and polarized LEDs with multi-view DSLR cameras, efficiently separating albedo and specular components and obtaining high-fidelity surface normals via polarization and gradient illuminations. [19] introduces second-order spherical gradient illumination for capturing specular roughness and anisotropy via a few captures. However, it’s limited to a single viewpoint due to linearly polarized illumination that requires precise tuning for albedo-specular separation. [20] further employs circularly polarized illumination to address view dependency, but fails to separate diffuse and specular reflectance or recover surface normals. In contrast, [43] proposes a comprehensive technique for measuring geometry and spatially-varying reflectance under continuous spherical harmonic illumination without a polarizer. While suitable for many objects, it fails with complex, non-convex objects with interreflection and occlusion. Real-world objects with varying materials and imperfect lighting and camera conditions result in artifacts like interreflection shadows, over-exposure, and lens glare, limiting the applicability of current capture methods.

In this paper, we present a novel, practical, and precise approach for acquiring spatially-varying reflectance and object geometry. We leverage a polarized reflectance field to densely capture objects from diverse lighting directions through three steps: 1) Data Preprocessing. We analyze, model, and reduce noise arising from uncontrollable factors like overexposure, inter-reflection, and lens flare. 2) Initialization. Utilizing the preprocessed imagery, we solve for initial albedo and specular separation under gradient illumination assumptions. 3) Optimization. This stage involves optimizing surface normals, anisotropy, and roughness while updating albedo and specular maps. Our experiments showcase significantly improved capture quality and accuracy. In summary, our contributions include:

  1. 1.

    A unique setup for capturing the polarized reflectance fields of an object.

  2. 2.

    A comprehensive solution for accurately measuring the geometry (surface normal) and spatially-varying reflectance of real-world objects, encompassing albedo, specular, roughness, and anisotropy parameters.

2 Related Works

2.1 Analytic Reflection Models

Ward [44] and the simplified Torrance-Sparrow [41] are widely used BRDF models in tasks to acquire reflectance field [38, 12]. While the simplified Torrance-Sparrow addresses isotropic reflection only, the Ward model, a simplified version of the Cook-Torrance model [10], is physically valid for both isotropic and anisotropic reflections. Additionally, BRDFs have been applied in recent physically based differentiable rendering techniques, often with certain approximations such as Blinn-Phong [32], isotropic specular reflection [8], cosine-weighted BRDF [49].

SVBRDFs encompass 2D maps of surface properties such as texture and roughness. Most studies focus on SVBRDF acquisition of planar surfaces [25, 33, 48, 13, 17, 36]. For non-planar objects, [34] predicted both shape and SVBRDF from a single image, but with limited photo-realism. Utilizing polarization cues under flash illumination, [14] achieved higher quality in specular effect but suffers from inaccurate diffuse albedo due to baked-in specular highlights. [26] captured the polarimetric SVBRDF, including the 3D Mueller matrix, yet lacked anisotropic specular effects. A recent work [18] captured both anisotropic reflectance at the microscopic level and employed an image translation network to propagate BRDF from micro to meso, successfully fitting specular reflectance without diffuse lobe influence. However, they did not decouple and explicitly optimize the specular parameters.

2.2 ML-based BTF capture

Recently, the bidirectional texture function (BTF) has been introduced to model finer reflection including mesoscopic effects such as subsurface scattering, interreflection, and self-occlusion across the surface [11]. Recent advancements utilize neural representations trained to replicate observations, as BTF lacks an analytical form. [45] synthesized BTF under different views and illuminations and trained an SVM classifier on the synthesized dataset to classify real-world materials based on a low-dimensional feature descriptor.  [40] trained an autoencoder that simulated discrete angular integration of the product of the reflectance signal with angular filters by projecting 4D light direction and RGB into a weighted matrix and then encoding them into a latent vector. The decoder outputs RGB from the latent vector concatenated with query directions. Upon this, [29] employed a neural texture pyramid instead of the encoder to represent multi-scale BTFs, achieving smaller storage but more levels of detail, i.e., accurate parallax and self-shadowing. [30] added surface curvature into BTF input and outputs opacity alongside RGB color. It also allowed for UV coordinate offset to handle silhouette and parallax effects for near-grazing viewing directions. [15] presented a biplane representation of BTF, including spatial and half feature maps, and employed a small universal MLP for radiance decoding, achieving a faster evaluation compared to the method [30]. Nonetheless, current BTF representations lack the flexibility and generality of SVBRDF. Aiming for the standard industrial rendering pipeline, we adopt SVBRDFs in this work. Although mesoscopic effects are beyond our focus, our approach does not contradict with any potential BTF extension.

2.3 Gradient Illumination for BRDF

We exclusively concentrate on polarized illumination, as it overtakes non-polarized approaches in acquiring the specular reflectance properties. [35] designed the methodology to separate diffuse and specular components and obtain corresponding normals from polarized 1st-order spherical gradient illumination patterns. While it assumes that the object is isotropic and has a small specular lobe throughout, [19] made a weaker specular BRDF assumption, only symmetry about the mean direction, and derived computation of roughness and anisotropy from the 2nd-order spherical gradient illumination. [20] separated and inferred the specular roughness from circularly polarized illumination using the Stokes vector parameters. [21] degraded the linear polarized pattern of [35] to two latitudinal and longitudinal patterns, allowing diffuse-specular separation for multiview stereo captures. The obtained photometric normals are then used to constrain further stereo reconstruction. [43] adopted up to 5th-order continuous spherical harmonic illumination to obtain diffuse, specular roughness and anisotropy. Instead of optimization, it used a table of a range of roughness and anisotropy to integrate the 5th-order spherical harmonics and found the best-matched specular parameters. With color spherical gradient illuminations and linear polarizers placed on cameras, [16] could acquire diffuse and specular albedo and normals simultaneously at a single shot with a Phong model. [23] addressed the unrealistic double shading issue in this single-shot approach using two different color gradient illuminations. [31] argued the complexity of using both color lights and cameras, adopting monochrome cameras that can still hallucinate parallel- and cross-polarized images under unpolarized illuminations. [27] adopted binary gradient illumination, which requires fewer photos than spherical gradient illumination. As discussed in Section 1, our capture is based on the linear polarized spherical gradient illumination [35].

3 Preliminary

Fresnel Equations

The Fresnel equations describe how the light behaves when it encounters the boundary between different optical media, involving two primary reflection components: specular and diffuse reflection. Specular reflection occurs unscattered, producing distinct reflections at any interface. In contrast, diffuse reflection results from both surface and subsurface scattering, causing light to scatter in various directions. The Fresnel equation specifies that specular reflection retains the polarization state of the incident light, while diffuse reflection remains unpolarized, regardless of the incident light’s polarization characteristics [35]. Therefore, diffuse and specular reflection can be separated with different states of polarization, which are determined by incident light conditions.

Linear Polarization and Malus’ Law

Theoretically, placing linear polarizers and analyzers in front of light sources and observers with different orientations can effectively separate diffuse and specular reflections. When the polarizer and analyzer are set perpendicular to each other, only the diffuse reflection becomes visible, and the intensity of polarized light is governed by Malus’s Law:

I=I0cos2θI=I_{0}\cos^{2}\theta (1)

where θ\theta is the angle between the polarizer’s axis and the analyzer’s axis. Given that the average of cos2θ\cos^{2}\theta is 12\frac{1}{2}, under identical lighting conditions, the radiant intensity of diffuse reflection IdI_{d} and specular reflection IsI_{s} can be measured by:

Id=2I,Is=2I2II_{d}=2I_{\perp},\quad I_{s}=2I_{\parallel}-2I_{\perp} (2)

where II_{\perp} and II_{\parallel} are the observations under cross and parallel-polarized lighting respectively; proof can be found in supplementary material via Mueller calculus.

Task Statement

Materials are represented using spatially-varying BRDF, which explains light scattering on a material’s surface in various directions. Our goal is to precisely measure the following reflectance attributes with controllable polarized lighting: diffuse and specular albedo ρd\rho_{d}, ρs\rho_{s}, diffuse and specular normal ndn_{d}, nsn_{s}, specular variance σ\sigma, anisotropy ς\varsigma, and roughness γ\gamma. We also measure diffuse and specular visibility νd\nu_{d}, νs\nu_{s}, inter-reflection ςd\varsigma_{d}, ςs\varsigma_{s}, and occlusion τd\tau_{d}, τs\tau_{s}.

Data Format

Table 1 summarizes the symbols and formats of the input and output data involved in our method. +\mathbb{R}_{+} encompasses all non-negative real numbers, and 𝔹={0,1}\mathbb{B}=\{0,1\}. HH and WW respectively refer to the height and width of the image. In our case, the capture is executed with 8 RED KOMODO 6K cameras at H=6144H=6144 and W=3240W=3240 at 30 FPS covering N=346N=346 lighting directions.

I/O Name Symbol Dimension
I Camera Pose 𝚁,𝚝,𝙺\mathtt{R},\mathtt{t},\mathtt{K} 3×3\mathbb{R}^{3\times 3}, 3\mathbb{R}^{3}, 3×3\mathbb{R}^{3\times 3}
I Captured Image II_{\perp}, II_{\parallel} H×W×3\mathbb{R}^{H\!\times\!W\!\times\!3}
I Captured OLAT Λ\Lambda_{\perp}, Λ\Lambda_{\perp} N×H×W×3\mathbb{R}^{N\!\times\!H\!\times\!W\!\times\!3}
O Visibility Map νd\nu_{d}, νs\nu_{s} 𝔹H×W\mathbb{B}^{H\!\times\!W}
O Occlusion Map τd\tau_{d}, τs\tau_{s} +H×W\mathbb{R}_{+}^{H\!\times\!W}
O Inter-reflection Map ϱd\varrho_{d}, ϱs\varrho_{s} +H×W×3\mathbb{R}_{+}^{H\!\times\!W\!\times\!3}, +H×W\mathbb{R}_{+}^{H\!\times\!W}
O Albedo Map ρd\rho_{d}, ρs\rho_{s} +H×W×3\mathbb{R}_{+}^{H\!\times\!W\!\times\!3}, +H×W\mathbb{R}_{+}^{H\!\times\!W}
O Normal Map ndn_{d}, nsn_{s} H×W×3\mathbb{R}^{H\!\times\!W\!\times\!3}
O Specular-var Map σ\sigma +H×W×2\mathbb{R}_{+}^{H\!\times\!W\times 2}
O Anisotropy Map ς\varsigma +H×W\mathbb{R}_{+}^{H\!\times\!W}
O Roughness Map γ\gamma +H×W\mathbb{R}_{+}^{H\!\times\!W}
Table 1: Symbols. Dimensions are indicated only once if the symbols with different subscripts are within the same dimension.

4 Methods

Overview

The complex interplay of diffuse and specular reflections in light transport challenges accurate material capture, often resulting in imprecise measurements. To address this issue, we conduct a thorough capture process and employ statistical and optimization methods on the recorded sequences for precise material acquisition.

Our method consists of three primary steps: We start with polarized OLAT (One Light at A Time) captures, encompassing cross and parallel polarization conditions (Section 4.1). Next (Section 4.2), we analyze and preprocess the captured sequence to eliminate overexposure. Additionally, we define a set of constraints aimed at reducing the influence of inter-reflection, self-occlusion, and lens flare during the subsequent optimization process. Finally (Section 4.3), we delve into the specifics of our optimization approach, dedicated to enhancing the accuracy of material properties.

Refer to caption
Figure 1: Polarized OLAT. We present an object captured under the first 300 lighting conditions. The top two rows exhibit images captured through cross-polarized OLAT and parallel-polarized OLAT. In the last row, we zoom in on details, marked by corresponding continuous/dotted boxes, with the active light board in the lower corner accordingly. Additionally, we present the lighting order, progressing from blue to red.

4.1 Polarized OLAT Capture

To separate diffuse and specular reflection, we perform the multiview captures of the same object under cross-polarized and parallel-polarized OLAT illuminations at the same intensity. The capture results in a cross-polarized sequence Λ={Ik}k=0N\Lambda_{\perp}=\{I_{\perp}^{k}\}_{k=0}^{N} and a parallel-polarized sequence Λ={Ik}k=0N\Lambda_{\parallel}=\{I_{\parallel}^{k}\}_{k=0}^{N} are shown in Fig. 1, where NN is the length of the sequence. Therefore, the diffuse reflection sequence Λd\Lambda_{d} and specular reflection sequence Λs\Lambda_{s} can be defined as:

Λd={2Ik}k=0N,Λs={2Ik2Ik}k=0N\Lambda_{d}=\{2I_{\perp}^{k}\}_{k=0}^{N},\quad\Lambda_{s}=\{2I_{\parallel}^{k}-2I_{\perp}^{k}\}_{k=0}^{N}\vspace{5pt} (3)

The transition of lighting polarization states is achieved by controlling the activation of different lights at each instance on the light board. On each light board, the white lights are arranged in a hexagonal pattern, with cross-polarizers and parallel polarizers placed alternately at the front. Throughout the capture process, each light of corresponding polarization states is activated via a 12-bit intensity code. Additionally, each OLAT sequence follows a spiral order from +z+z to z-z covering all available lighting directions ωi\omega_{i} over the sphere, in total NN directions. The direction of outgoing radiance ωo\omega_{o} is determined by the camera pose [𝚁,𝚝,𝙺][\mathtt{R},\mathtt{t},\mathtt{K}] from multiview camera calibration.

4.2 Analysis and Preprocess

While analyzing images under OLAT illumination, material observation may encompass overexposure highlights, inter-reflection, and self-occlusion. These physical phenomena can introduce inaccuracies in material measurement. In the subsequent paragraph, we visually elaborate on these effects and explain how we mitigate their impact.

Refer to caption
Figure 2: Overexposure Removal. We demonstrate the effectiveness of overexposure elimination by using a mirrorball. The observation under all-white lighting is a summation of all frames from the cross-polarized OLAT data Λ\Lambda_{\perp}. We select points a and c from the zoom-in region and b from the base, which is made of relatively diffuse material. The second row presents the corresponding intensities recorded under cross-polarized OLAT lighting. The horizontal axis represents the OLAT index, while the vertical axis indicates the recorded intensities in red, green, and blue.
Algorithm 1 Overexposure Removal
iter0;M2;δmean(Λ)\text{iter}\leftarrow 0;M\leftarrow 2;\delta\leftarrow\text{mean}(\Lambda)
while iter<M\text{iter}<M do
    for each pixel (i,j)(i,j) do
         λΛ(i,j)\lambda\leftarrow\Lambda^{(i,j)}
         iddescargsort(λ)\text{id}_{\text{desc}}\leftarrow\operatorname*{argsort}(\lambda)
         Δdescλ[iddesc[0:end1]]λ[iddesc[1:end]]\Delta_{\text{desc}}\leftarrow\lambda[\text{id}_{\text{desc}}[0:\text{end}\!-\!1]]-\lambda[\text{id}_{\text{desc}}[1:\text{end}]]
         id>ε=argwhere(Δdesc>ε)\text{id}_{>\varepsilon}=\operatorname*{argwhere}(\Delta_{\text{desc}}>\varepsilon)
         λ[id>ε[0]]λ[iddesc[1]]+δ\lambda[\text{id}_{>\varepsilon}[0]]\leftarrow\lambda[\text{id}_{\text{desc}}[1]]+\delta
    end for
    iteriter+1\text{iter}\leftarrow\text{iter}+1
end while

Overexposure

Overexposure occurs when intense light interacts with a material, producing highlights that mimic the pattern of the light sources.

For a particular surface point, the intensity observation should exhibit continuity within a specific range rather than displaying an emergent highlight. Exploiting this property allows for the immediate identification of abnormal pulses indicative of overexposure. As shown in Fig. 2, the intensity variations of three surface points under OLAT illumination clearly reveal the recognition of overexposure through strong pulses in the sequence.

Our approach, formulated in Algorithm 1, consists of sorting the signal intensities in each color channel according to the lighting order, detecting differences that exceed a predefined threshold ε\varepsilon, and replacing the values at such points with an ambient value δε\delta\ll\varepsilon. Usually, the threshold ε\varepsilon relates to the light sources and can be easily determined during the capture. The captured data at each pixel position (i,j)(i,j) within the OLAT sequence Λ{Λd,Λs}\Lambda\in\{\Lambda_{d},\Lambda_{s}\} is treated as a time-domain signal λ=Λ(i,j)N×3\lambda=\Lambda^{(i,j)}\in\mathbb{R}^{N\times 3} and this procedure can be iterated multiple times MM for a clean result.

As a result, overexposure can be effectively eliminated, yielding the calibrated OLAT sequence Λ^{Λ^d,Λ^s}\hat{\Lambda}\in\{\hat{\Lambda}_{d},\hat{\Lambda}_{s}\}. As depicted in Fig. 2, our method successfully isolates overexposure from the original capture.

Refer to caption
Figure 3: Inter-reflection and Lens Flare. In row 1), we present the captured intensity distribution of a fixed surface point, indicated with a cross (×\mathbf{\times}) in the following example images. The intensity patterns are identified as a) interreflection, b) regular specular reflection, and c) lens flare. Row 2) provides a false-color view, with a zoom-in view in the last row 3), as well as the raw capture. In the false-color view, the intensity strength is represented by the color, with stronger intensities appearing redder.

Inter-reflection

Moreover, the intricate behavior of light as it bounces around often gives rise to inter-reflection, particularly in the presence of highly specular objects in the environment. This can introduce additional errors when measuring material properties. Inside the capture device, inter-reflection predominantly occurs from the reflection of light sources bouncing off the capturing layout and being captured by the camera as depicted in Fig. 3. Furthermore, objects with concave geometries tend to exhibit a higher incidence of inter-reflection.

When the lighting arises from the lower hemisphere Ω\Omega^{-}, the object becomes unobservable due to the absence of incoming radiance. However, such a point can still be captured in the sequence caused by inter-reflection from ωir\omega_{ir}. This phenomenon typically occurs on the opposite side of the direction of the active lights ωi\omega_{i}, leading to ωirωi<0\omega_{ir}\cdot\omega_{i}<0. Usually, the surface point is observable when nωi>0n\cdot\omega_{i}>0. By introducing the visibility ν\nu, the inter-reflection can be approximated via:

ϱ=Ων(ωi,ωo)R(ωi,n)𝑑ωik=0NI^kζmax(ωikn,0)I^k\begin{split}\varrho&=\int_{\Omega^{-}}\nu(\omega_{i},\omega_{o})R(\omega_{i},n)d\omega_{i}\\ &\approx\sum\nolimits_{k=0}^{N}\lceil\hat{I}^{k}-\zeta\rceil\cdot max(-\omega_{i}^{k}\cdot n,0)\cdot\hat{I}^{k}\end{split} (4)

where ν(ωi,ωo)\nu(\omega_{i},\omega_{o}) is the visibility of ωi\omega_{i} observed via ωo\omega_{o}, R(ωir,n)R(\omega_{ir},n) is the surface reflection function, ζ\zeta is the average ambient noise map, \lceil\cdot\rceil is the ceiling operator, nn is the surface normal and I^k\hat{I}^{k} is the k-th image from the calibrated OLAT sequence Λ^{Λ^d,Λ^s}\hat{\Lambda}\in\{\hat{\Lambda}_{d},\hat{\Lambda}_{s}\}, where different polarized OLAT yields diffuse inter-reflection ϱd\varrho_{d} and specular inter-reflection ϱs\varrho_{s} accordingly. Given such observations, this inter-reflection can be mitigated by simply imposing a constraint on the incident lighting at each optimization step, specifically nωi0n\cdot\omega_{i}\geq 0.

Refer to caption
Figure 4: Acquisition. We present a measured mug comprising a diffuse base and a clear coat showcasing a1) original, b1) diffuse albedo ρd\rho_{d}, c1) specular albedo ρs\rho_{s}, a2) diffuse normal ndn_{d}, b2) diffuse inter-reflection ϱd\varrho_{d} (intensity adjusted for better visualization), c2) diffuse occlusion τd\tau_{d}, a3) specular normal nsn_{s}, b3) specular inter-reflection ϱs\varrho_{s}, c3) specular occlusion τd\tau_{d}, a4) specular variance σ\sigma, b4) anisotropy ς\varsigma, and c4) roughness γ\gamma.

Lens Flare

Additionally, lens flare arises when the lighting direction aligns with the opposite side of the viewing angle ωo\omega_{o} as shown in Fig. 3. This condition yields in ωiωo<0\omega_{i}\cdot\omega_{o}<0 and, as a result, nωi<0n\cdot\omega_{i}<0 when the surface point is observable via ωo\omega_{o}. The effect of lens flare can therefore be reduced with the constraint nωi0n\cdot\omega_{i}\geq 0.

Occlusion

Another phenomenon that can impact material acquisition is self-occlusion, leading to the presence of shadows in the observations. The more shadows appear in the observations, the less accurate the material measurements become. The occlusion, τ\tau, is defined via:

τ=Ων(ωi,ωo)(ωin)𝑑ωi(Ω(ωin)𝑑ωi)14Nk=0NI^kζmax(ωikn,0)\begin{split}\tau&=\int_{\Omega}\nu(\omega_{i},\omega_{o})\cdot(\omega_{i}\cdot n)d\omega_{i}\cdot\left(\int_{\Omega}(\omega_{i}\cdot n)d\omega_{i}\right)^{-1}\\ &\approx\frac{4}{N}\sum\nolimits_{k=0}^{N}\lceil\hat{I}^{k}-\zeta\rceil\cdot max(\omega_{i}^{k}\cdot n,0)\end{split} (5)

where ωi\omega_{i}, ωo\omega_{o}, nn, ν(ωi,ωo)\nu(\omega_{i},\omega_{o}), ζ\zeta, and I^k\hat{I}^{k} are the same as in Equation 4. 4/N4/N is a factor to normalize the average occlusion in the hemisphere (discussed in supplementary material). The measured occlusion map can be used further to eliminate shadows from the albedo map.

Refer to caption
Figure 5: Normal Fusion. a1) original image, b1) maximum cross-correlation of diffuse normal optimization, c1) maximum cross-correlation of specular normal optimization, a2) fused normal, b2) diffuse normal, c2) specular normal.

4.3 Optimization

Following the preprocess, our focus shifts to solving optimization problems to obtain the required material. The process begins with an initial approximation of the solution, achieved through the synthesized gradient illumination. This is followed by successive refinements of surface normals nn, anisotropy ς\varsigma, roughness γ\gamma, and albedo ρ\rho, each step methodically integrating constraints that emerge from our analytical evaluations. An illustration of the acquired material and intermediate outcomes is presented in Fig. 4. Full derivation can be found in supplementary material.

Initialize ρ\rho and nn

Initial diffuse albedo ρdinit\rho_{d}^{\text{init}} and specular albedo ρsinit\rho_{s}^{\text{init}} can be easily derived from the preprocessed sequences Λ^d\hat{\Lambda}_{d} and Λ^s\hat{\Lambda}_{s} via:

ρdinit=ΩRd(ωi,ωo)𝑑ωi4κNk=0NI^dk\rho_{d}^{\text{init}}=\int_{\Omega}R_{d}(\omega_{i},\omega_{o})d\omega_{i}\approx\frac{4\kappa}{N}\sum\nolimits_{k=0}^{N}\hat{I}_{d}^{k} (6)
ρsinit=ΩRs(ωi,ωo)𝑑ωi4πκNk=0NI^sk\rho_{s}^{\text{init}}=\int_{\Omega}R_{s}(\omega_{i},\omega_{o})d\omega_{i}\approx\frac{4\pi\kappa}{N}\sum\nolimits_{k=0}^{N}\hat{I}_{s}^{k} (7)

where RdR_{d} and RsR_{s} represent the diffuse reflection function and the specular reflection function of the material, respectively. κ\kappa is a constant determined by light intensity and solid angle in the capture device.

Furthermore, the initial estimates for the diffuse surface normal ndinitn_{d}^{\text{init}} and specular normal nsinitn_{s}^{\text{init}} can be approximated from the spherical gradient illumination pattern Pj{Px,Py,Pz}P_{j}\in\{P_{x},P_{y},P_{z}\} [35]. The response under gradient lighting pattern, including negative values, I^d,j{I^d,x,I^d,y,I^d,z}\hat{I}_{d,j}\in\{\hat{I}_{d,x},\hat{I}_{d,y},\hat{I}_{d,z}\} and I^s,j{I^s,x,I^s,y,I^s,z}\hat{I}_{s,j}\in\{\hat{I}_{s,x},\hat{I}_{s,y},\hat{I}_{s,z}\} can be easily synthesized using the captured OLAT sequence with weighting w[1,1]w\in[-1,1] over the incident lighting direction. The derivation and synthesized results can be found in the supplementary material.

I^d,j=ΩPj(ωi)Rd(ωi,ωo)𝑑ωik=0NwjkI^dk\hat{I}_{d,j}=\int_{\Omega}\!P_{j}(\omega_{i})R_{d}(\omega_{i},\omega_{o})d\omega_{i}\approx\sum\nolimits_{k=0}^{N}w_{j}^{k}\hat{I}_{d}^{k} (8)
I^s,j=ΩPj(ωi)Rs(ωi,ωo)𝑑ωik=0NwjkI^sk\hat{I}_{s,j}=\int_{\Omega}\!P_{j}(\omega_{i})R_{s}(\omega_{i},\omega_{o})d\omega_{i}\approx\sum\nolimits_{k=0}^{N}w_{j}^{k}\hat{I}_{s}^{k} (9)

Moreover, with 𝔑()\mathfrak{N}(\cdot) representing the normalization operator, the surface normal can be derived as follows:

ndinit=𝔑(32πρd[I^d,x,I^d,y,I^d,z])nsinit=𝔑(𝔑[I^s,x,I^s,y,I^s,z])+ωo)\begin{split}n_{d}^{\text{init}}&=\mathfrak{N}\left(\frac{3}{2\pi\rho_{d}}\cdot[\hat{I}_{d,x},\hat{I}_{d,y},\hat{I}_{d,z}]\right)\\ n_{s}^{\text{init}}&=\mathfrak{N}\big{(}\mathfrak{N}[\hat{I}_{s,x},\hat{I}_{s,y},\hat{I}_{s,z}])+\omega_{o}\big{)}\\ \end{split} (10)
Refer to caption
Figure 6: Ablation study on overexposure removal. The proposed overexposure removal effectively mitigates the lighting baked-in effect from the acquired specular albedo.
Refer to caption
Figure 7: Ablation study on optimization. Without optimization, the acquired diffuse normal incorrectly reflects texture as normal. This shows much better separation after our proposed optimization step.

Refinement on nn

Λ^d\hat{\Lambda}_{d} captures the diffuse characteristics of surface points under periodic illumination. The recorded data represents the area element on the object, dAdA, observed through an aperture of area dAodA_{o}, thus subtending a solid angle dΩod\Omega_{o}. For any arbitrary equal angle dΩd\Omega, each surface point within Λ^d\hat{\Lambda}_{d} is expected to exhibit Lambertian appearance in accordance with Lambert’s law:

Lo=Li(ωin)(ωon)dΩdAdΩo(ωon)dAo=Li(ωin)dΩdAdΩodAo\begin{split}L_{o}=\frac{L_{i}(\omega_{i}\cdot n)(\omega_{o}\cdot n)d\Omega dA}{d\Omega_{o}(\omega_{o}\cdot n)dA_{o}}=\frac{L_{i}(\omega_{i}\cdot n)d\Omega dA}{d\Omega_{o}dA_{o}}\end{split} (11)

where LiL_{i} is the incident radiance from the light source and LoL_{o} denotes the observed radiance. By maintaining uniformity in LiL_{i} across the reflection sequence Λ^d\hat{\Lambda}_{d}, the observation of a surface point is solely influenced by the incident lighting direction ωi\omega_{i} and the surface normal nn. Consequently, we can get optimized normal n^d\hat{n}_{d} by maximizing the overall cross-correlation with the observation Λ^d\hat{\Lambda}_{d} at each surface point through the expression:

n^d=argmaxnninit:=ndinit𝔑([νd0nωi0νdknωik])T𝔑([I^d0I^dk])s.t.n=1,nωik>0, and I^dkΛ^d\begin{gathered}\!\hat{n}_{d}\!=\!\operatorname*{argmax}_{\begin{subarray}{c}n\\ n^{\text{init}}:=n_{d}^{\text{init}}\end{subarray}}\mathfrak{N}\left(\!\begin{bmatrix}\nu_{d}^{0}\!\cdot\!n\!\cdot\!\omega_{i}^{0}\\ \cdots\\ \nu_{d}^{k}\!\cdot\!n\!\cdot\!\omega_{i}^{k}\end{bmatrix}\!\right)^{\!T\!}\!\mathfrak{N}\left(\!\begin{bmatrix}\hat{I}_{d}^{0}\\ \cdots\\ \hat{I}_{d}^{k}\end{bmatrix}\!\right)\\ \text{s.t.}\quad\lVert n\rVert=1,n\cdot\omega_{i}^{k}>0\text{, and\,}\hat{I}_{d}^{k}\in\hat{\Lambda}_{d}\end{gathered} (12)

Likewise, the refined specular normal n^s\hat{n}_{s} can be attained by considering the cross-correlation between the reflection ωrk=2(ωikn)nωik\omega_{r}^{k}=2(\omega_{i}^{k}\cdot n)n-\omega_{i}^{k} and the observation:

n^s=argmaxnninit:=nsinit𝔑([νs0ωr0ωoνskωrkωo])T𝔑([I^s0I^sk])s.t.n=1,nωik>0, and I^skΛ^s\begin{gathered}\hat{n}_{s}\!=\!\operatorname*{argmax}_{\begin{subarray}{c}n\\ n^{\text{init}}:=n_{s}^{\text{init}}\end{subarray}}\mathfrak{N}\left(\!\begin{bmatrix}\!\nu_{s}^{0}\!\cdot\!\omega_{r}^{0}\!\cdot\!\omega_{o}\!\\ \cdots\\ \!\nu_{s}^{k}\!\cdot\!\omega_{r}^{k}\!\cdot\!\omega_{o}\!\end{bmatrix}\!\right)^{\!T\!}\!\mathfrak{N}\left(\!\begin{bmatrix}\hat{I}_{s}^{0}\\ \cdots\\ \hat{I}_{s}^{k}\end{bmatrix}\!\right)\\ \text{s.t.}\quad\lVert n\rVert=1,n\cdot\omega_{i}^{k}>0\text{, and\,}\hat{I}_{s}^{k}\in\hat{\Lambda}_{s}\end{gathered} (13)

In most cases, the surface normals obtained through cross-polarized OLAT and parallel-polarized OLAT tend to exhibit similarity. However, in cases where the material exhibits stronger energy absorption, often indicated by a darker appearance, the diffuse reflection weakens, leading to inaccuracies in the diffuse normal. Also, when dealing with materials of a more intricate structure, inter-reflections and self-occlusion occur more frequently during the capture process, resulting in inaccurate specular normals.

The optimization in Equations 12 and 13 assesses the alignment between normals and observations, allowing us to further improve normal quality. These cross-correlation coefficients, forming vectors in 2\mathbb{R}^{2} for each pixel, are normalized and serve as blending weights for enhancing the measured normals, as shown in Fig. 5.

Refer to caption
Figure 8: Anisotropy Analysis. We showcase a) the original capture, b) the anisotropy map ς\mathbf{\varsigma}, and c) the anisotropy variation along both vertical and horizontal directions as indicated by colored arrows on two objects. In the chart, the horizontal axis represents the pixel index, while the vertical axis is the measured anisotropy.

Anisotropy ς\varsigma and Roughness γ\gamma

Via captured OLAT, calculating the shape of the reflection lobe becomes straightforward, enabling the measurement of material isotropy/anisotropy. Ideally, the specular lobe follows a normal distribution governed by σ=(σx,σy)\sigma=(\sigma_{x},\sigma_{y}), defined as [44]:

fσ(ωi,ωo,n)=14πσxσy(ωon)(ωin)exp(2htσx2+hbσy21+hn)f_{\sigma}(\omega_{i},\omega_{o},n)\!=\!\frac{1}{4\pi\sigma_{x}\sigma_{y}\sqrt{(\omega_{o}\!\cdot\!n)(\omega_{i}\!\cdot\!n)}}\exp\!{\left(\!-2\frac{\frac{h\cdot t}{\sigma_{x}}^{\!2\!}\!+\!\frac{h\cdot b}{\sigma_{y}}^{2}}{1\!+\!h\!\cdot\!n}\right)}

where h=𝔑(ωi+ωo)h=\mathfrak{N}(\omega_{i}+\omega_{o}) is the halfway vector, [n,b,t][n,b,t] defines the local shading frame that aligns to the optimized specular surface normal n^s\hat{n}_{s}. Following the collection of the response Λ^\hat{\Lambda}, the optimization of the variance σ\sigma of specular reflection, aimed at achieving the closest match to the observation, can be performed as:

σ^=argminσ+2𝔑([fσ(ωi0,ωo,n^s)fσ(ωik,ωo,n^s)])𝔑([I^s0I^sk])2s.t.n^sωik>0, and I^skΛ^s\begin{gathered}\hat{\sigma}\!=\!\operatorname*{argmin}_{\sigma\in\mathbb{R}_{+}^{2}}\!\left\lVert\mathfrak{N}\!\left(\!\begin{bmatrix}\!f_{\sigma}(\omega_{i}^{0},\omega_{o},\hat{n}_{s})\!\\ \cdots\\ \!f_{\sigma}(\omega_{i}^{k},\omega_{o},\hat{n}_{s})\!\end{bmatrix}\!\right)\!-\!\mathfrak{N}\!\left(\!\begin{bmatrix}\hat{I}_{s}^{0}\\ \cdots\\ \hat{I}_{s}^{k}\end{bmatrix}\!\right)\!\right\rVert^{2}\\ \text{s.t.}\quad\hat{n}_{s}\cdot\omega_{i}^{k}>0\text{, and\,}\hat{I}_{s}^{k}\in\hat{\Lambda}_{s}\end{gathered} (14)

When the material is isotropic, the specular lobe is symmetric, where σxσy\sigma_{x}\!\approx\!\sigma_{y}. In such cases, diffuse material tends to have a flat and wide reflection lobe while that specular material is narrow and sharp. Material anisotropy ς\varsigma and material roughness γ\gamma can therefore be derived by:

ς=σxσyσx+σy,γ=σx2+σy2\varsigma=\frac{\sigma_{x}-\sigma_{y}}{\sigma_{x}+\sigma_{y}},\quad\gamma=\lVert\sigma_{x}^{2}+\sigma_{y}^{2}\rVert (15)

Refinement on ρ\rho

The refined diffuse normal n^d\hat{n}_{d} can be further employed to measure the diffuse albedo ρ^d\hat{\rho}_{d} through:

ρ^d=argminρρinit:=ρdinitρ[n^dωi0n^dωik][I^d0I^dk]2s.t.n^dωik>0, and I^dkΛ^d\begin{gathered}\hat{\rho}_{d}=\operatorname*{argmin}_{\begin{subarray}{c}\rho\\ \rho^{\text{init}}:=\rho_{d}^{\text{init}}\end{subarray}}\!\left\lVert\rho\!\cdot\!\begin{bmatrix}\hat{n}_{d}\!\cdot\!\omega_{i}^{0}\\ \cdots\\ \hat{n}_{d}\!\cdot\!\omega_{i}^{k}\end{bmatrix}-\!\begin{bmatrix}\hat{I}_{d}^{0}\\ \cdots\\ \hat{I}_{d}^{k}\end{bmatrix}\!\right\lVert^{2}\\ \text{s.t.}\quad\hat{n}_{d}\cdot\omega_{i}^{k}>0\text{, and\,}\hat{I}_{d}^{k}\in\hat{\Lambda}_{d}\end{gathered} (16)

Also, the specular albedo can be optimized similarly via the optimized specular normal n^s\hat{n}_{s}, and optimized variance σ^\hat{\sigma}:

ρ^s=argminρρinit:=ρsinitρ[fσ^(ωi0,ωo,n^s)fσ^(ωik,ωo,n^s)][I^s0I^sk]2s.t.n^sωik>0, and I^skΛ^s\begin{gathered}\hat{\rho}_{s}\!=\!\operatorname*{argmin}_{\begin{subarray}{c}\rho\\ \rho^{\text{init}}:=\rho_{s}^{\text{init}}\end{subarray}}\!\left\lVert\rho\!\cdot\!\begin{bmatrix}\!f_{\hat{\sigma}}(\omega_{i}^{0},\omega_{o},\hat{n}_{s})\!\\ \cdots\\ \!f_{\hat{\sigma}}(\omega_{i}^{k},\omega_{o},\hat{n}_{s})\!\end{bmatrix}-\!\begin{bmatrix}\hat{I}_{s}^{0}\\ \cdots\\ \hat{I}_{s}^{k}\end{bmatrix}\!\right\rVert^{2}\\ \text{s.t.}\quad\hat{n}_{s}\cdot\omega_{i}^{k}>0\text{, and\,}\hat{I}_{s}^{k}\in\hat{\Lambda}_{s}\end{gathered} (17)

Additionally, the presence of shadows in the albedo can be partially mitigated by factoring in occlusion through ρ^/τ\hat{\rho}/\tau, where ρ^{ρ^d,ρ^s},τ{τd,τs}\hat{\rho}\!\in\!\{\hat{\rho}_{d},\hat{\rho}_{s}\},\tau\!\in\!\{\tau_{d},\tau_{s}\}. However, it’s important to acknowledge that ρ^/τ\hat{\rho}/\tau is vulnerable to noise.

5 Results and Experiments

Refer to caption
Figure 9: Rendering Comparisons. We validate our results using physically-based renderings with the measured materials. For each object, we showcase the renderings with corresponding captures under 1) environmental lighting or 2) area lighting conditions.
Refer to caption
Figure 10: Qualitative Comparison. We compare results from 1) our method with 2) from [35] via static capture on objects with specular outer layers. Examined properties cover diffuse albedo ρd\rho_{d} and specular albedo ρs\rho_{s}, diffuse normal ndn_{d}, and specular normal nsn_{s}, with zoomed-in views. Normally, diffuse and specular normals are similar, but in multi-layered materials, they may differ slightly.
Refer to caption

Figure 11: Reference Images for Figure 10. These are original captures with all-white lights on.

Ablation Study

In Fig. 6 and Fig. 7, we highlight the importance of our proposed methods by showcasing the difference between utilizing or omitting the overexposure removal module and optimization. These steps are pivotal throughout the entire process: the removal of overexposure aids in cleanly separating diffuse and specular components, while optimized physically correct normals further enhance the quality of acquired material components afterward.

Anisotropy Distribution Analysis

We visualize the measured anisotropy values ς\varsigma along specific directions in Fig. 8. For diffuse objects, the measured anisotropy across different directions is nearly zero, indicating that the reflection lobe is evenly distributed, demonstrating isotropic behavior. Conversely, for the specular object, the measured value tends to deviate, indicating anisotropy. In contrast, for specular objects, the measured values deviate, indicating anisotropy. For the mug, the red scanline crosses both the clearcoat and diffuse base, clearly distinguishing the two materials. The anisotropy variation along the scanline shows distinct splits, confirming our method’s correctness. Furthermore, the renderings in Fig. 9 of metal soda also validate this. Without anisotropy, the specular reflections on the cylindrical object would result in a spherical specular highlight rather than a linear strip.

Relighting

In Fig. 9, we showcase image-based rendering achieved in Blender, using our measured material properties and various HDRI as global illumination. The reference image for HDRI illumination is synthesized by weighting the captured OLAT images for each light, with these weights derived from averaging the pixel values within corresponding spherical areas of the HDRI light probe. In Blender, we use the principal BSDF material, incorporating the measured diffuse albedo as the base color, the diffuse normal as the object’s normal, and the measured specular albedo to set the specular IOR level along with the measured tangent. The measured roughness is also applied. Under HDRI lighting, our renderings closely match the reference image, and under area lighting conditions, our method accurately captures specular reflections on highly glossy materials.

Qualitative Comparisons

We compare our results with [35], using synthesized gradient illumination from polarized OLAT capture, and static gradient illumination capture. The latter, designed for human skin, incorporates lighting pattern adjustments and lowers the lighting intensity to avoid lens flare and unexpected highlights, as shown in Fig. 10 with references in Fig. 11. Due to space limits, more results can be found in the supplementary.

Ma et al. [35] struggle to eliminate overexposure on object surfaces. Moreover, the mixture of specular reflection with diffuse reflection can compromise the quality of albedo. This may further affect the accuracy of captured normals when albedo is introduced in Equation 10. In contrast, our method achieves a distinct separation between the diffuse albedo and the specular albedo, effectively reducing overexposures on both maps and leading to a more accurate reflectance measurement. Furthermore, our approach enhances the captured diffuse normals and specular components while mitigating inaccuracies arising from albedo maps, accurately preserving intricate geometric details.

6 Conclusion

In this work, we introduce the polarized reflectance field for precise material acquisition. Our results showcase a comprehensive enhancement across various material layers through preprocessing and optimization, in alignment with physical principles. Nevertheless, certain challenges persist. System inaccuracies arise from severe inter-reflections, involving the intricate distinction between direct and inter-reflected light. More limitations are detailed in the supplementary material. These challenges and limitations could potentially be addressed with a neural network in the future.

References

  • Barron and Malik [2014] Jonathan T Barron and Jitendra Malik. Shape, illumination, and reflectance from shading. IEEE transactions on pattern analysis and machine intelligence, 37(8):1670–1687, 2014.
  • Basri et al. [2007] Ronen Basri, David Jacobs, and Ira Kemelmacher. Photometric stereo with general, unknown lighting. International Journal of computer vision, 72:239–257, 2007.
  • Bazaraa et al. [2013] Mokhtar S Bazaraa, Hanif D Sherali, and Chitharanjan M Shetty. Nonlinear programming: theory and algorithms. John wiley & sons, 2013.
  • Blondel et al. [2021] Mathieu Blondel, Quentin Berthet, Marco Cuturi, Roy Frostig, Stephan Hoyer, Felipe Llinares-López, Fabian Pedregosa, and Jean-Philippe Vert. Efficient and modular implicit differentiation. arXiv preprint arXiv:2105.15183, 2021.
  • Boss et al. [2021a] Mark Boss, Raphael Braun, Varun Jampani, Jonathan T Barron, Ce Liu, and Hendrik Lensch. Nerd: Neural reflectance decomposition from image collections. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 12684–12694, 2021a.
  • Boss et al. [2021b] Mark Boss, Varun Jampani, Raphael Braun, Ce Liu, Jonathan Barron, and Hendrik Lensch. Neural-pil: Neural pre-integrated lighting for reflectance decomposition. Advances in Neural Information Processing Systems, 34:10691–10704, 2021b.
  • Bradbury et al. [2018] James Bradbury, Roy Frostig, Peter Hawkins, Matthew James Johnson, Chris Leary, Dougal Maclaurin, George Necula, Adam Paszke, Jake VanderPlas, Skye Wanderman-Milne, and Qiao Zhang. JAX: composable transformations of Python+NumPy programs, 2018.
  • Cai et al. [2022] Guangyan Cai, Kai Yan, Zhao Dong, Ioannis Gkioulekas, and Shuang Zhao. Physics-based inverse rendering using combined implicit and explicit geometries. In Computer Graphics Forum, pages 129–138. Wiley Online Library, 2022.
  • Collett [2005] Edward Collett. Field guide to polarization. Spie Bellingham, WA, 2005.
  • Cook and Torrance [1982] Robert L Cook and Kenneth E. Torrance. A reflectance model for computer graphics. ACM Transactions on Graphics (ToG), 1(1):7–24, 1982.
  • Dana et al. [1999] Kristin J Dana, Bram Van Ginneken, Shree K Nayar, and Jan J Koenderink. Reflectance and texture of real-world surfaces. ACM Transactions On Graphics (TOG), 18(1):1–34, 1999.
  • Debevec et al. [2000] Paul Debevec, Tim Hawkins, Chris Tchou, Haarm-Pieter Duiker, Westley Sarokin, and Mark Sagar. Acquiring the reflectance field of a human face. In Proceedings of the 27th annual conference on Computer graphics and interactive techniques, pages 145–156, 2000.
  • Deschaintre et al. [2018] Valentin Deschaintre, Miika Aittala, Fredo Durand, George Drettakis, and Adrien Bousseau. Single-image svbrdf capture with a rendering-aware deep network. ACM Transactions on Graphics (ToG), 37(4):1–15, 2018.
  • Deschaintre et al. [2021] Valentin Deschaintre, Yiming Lin, and Abhijeet Ghosh. Deep polarization imaging for 3d shape and svbrdf acquisition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15567–15576, 2021.
  • Fan et al. [2023] Jiahui Fan, Beibei Wang, Milos Hasan, Jian Yang, and Ling-Qi Yan. Neural biplane representation for btf rendering and acquisition. In ACM SIGGRAPH 2023 Conference Proceedings, pages 1–11, 2023.
  • Fyffe and Debevec [2015] Graham Fyffe and Paul Debevec. Single-shot reflectance measurement from polarized color gradient illumination. In 2015 IEEE International Conference on Computational Photography (ICCP), pages 1–10. IEEE, 2015.
  • Gao et al. [2019] Duan Gao, Xiao Li, Yue Dong, Pieter Peers, Kun Xu, and Xin Tong. Deep inverse rendering for high-resolution svbrdf estimation from an arbitrary number of images. ACM Trans. Graph., 38(4):134–1, 2019.
  • Garces et al. [2023] Elena Garces, Victor Arellano, Carlos Rodriguez-Pardo, David Pascual-Hernandez, Sergio Suja, and Jorge Lopez-Moreno. Towards material digitization with a dual-scale optical system. ACM Transactions on Graphics (TOG), 42(4):1–13, 2023.
  • Ghosh et al. [2009] Abhijeet Ghosh, Tongbo Chen, Pieter Peers, Cyrus A Wilson, and Paul Debevec. Estimating specular roughness and anisotropy from second order spherical gradient illumination. In Computer Graphics Forum, pages 1161–1170. Wiley Online Library, 2009.
  • Ghosh et al. [2010] Abhijeet Ghosh, Tongbo Chen, Pieter Peers, Cyrus A Wilson, and Paul Debevec. Circularly polarized spherical illumination reflectometry. In ACM SIGGRAPH Asia 2010 papers, pages 1–12. 2010.
  • Ghosh et al. [2011] Abhijeet Ghosh, Graham Fyffe, Borom Tunwattanapong, Jay Busch, Xueming Yu, and Paul Debevec. Multiview face capture using polarized spherical gradient illumination. ACM Transactions on Graphics (TOG), 30(6):1–10, 2011.
  • Goldman et al. [2009] Dan B Goldman, Brian Curless, Aaron Hertzmann, and Steven M Seitz. Shape and spatially-varying brdfs from photometric stereo. IEEE Transactions on Pattern Analysis and Machine Intelligence, 32(6):1060–1071, 2009.
  • Guo et al. [2019] Kaiwen Guo, Peter Lincoln, Philip Davidson, Jay Busch, Xueming Yu, Matt Whalen, Geoff Harvey, Sergio Orts-Escolano, Rohit Pandey, Jason Dourgarian, et al. The relightables: Volumetric performance capture of humans with realistic relighting. ACM Transactions on Graphics (ToG), 38(6):1–19, 2019.
  • Hager and Zhang [2005] William W Hager and Hongchao Zhang. A new conjugate gradient method with guaranteed descent and an efficient line search. SIAM Journal on optimization, 16(1):170–192, 2005.
  • Hui et al. [2017] Zhuo Hui, Kalyan Sunkavalli, Joon-Young Lee, Sunil Hadap, Jian Wang, and Aswin C Sankaranarayanan. Reflectance capture using univariate sampling of brdfs. In Proceedings of the IEEE International Conference on Computer Vision, pages 5362–5370, 2017.
  • Hwang et al. [2022] Inseung Hwang, Daniel S Jeon, Adolfo Munoz, Diego Gutierrez, Xin Tong, and Min H Kim. Sparse ellipsometry: portable acquisition of polarimetric svbrdf and shape with unstructured flash photography. ACM Transactions on Graphics (TOG), 41(4):1–14, 2022.
  • Kampouris et al. [2018] Christos Kampouris, Stefanos Zafeiriou, and Abhijeet Ghosh. Diffuse-specular separation using binary spherical gradient illumination. In EGSR (EI&I), pages 1–10, 2018.
  • Kerbl et al. [2023] Bernhard Kerbl, Georgios Kopanas, Thomas Leimkühler, and George Drettakis. 3d gaussian splatting for real-time radiance field rendering. ACM Transactions on Graphics, 42(4), 2023.
  • Kuznetsov [2021] Alexandr Kuznetsov. Neumip: Multi-resolution neural materials. ACM Transactions on Graphics (TOG), 40(4), 2021.
  • Kuznetsov et al. [2022] Alexandr Kuznetsov, Xuezheng Wang, Krishna Mullia, Fujun Luan, Zexiang Xu, Milos Hasan, and Ravi Ramamoorthi. Rendering neural materials on curved surfaces. In ACM SIGGRAPH 2022 conference proceedings, pages 1–9, 2022.
  • LeGendre et al. [2018] Chloe LeGendre, Kalle Bladin, Bipin Kishore, Xinglei Ren, Xueming Yu, and Paul Debevec. Efficient multispectral facial capture with monochrome cameras. In ACM SIGGRAPH 2018 Posters, pages 1–2. 2018.
  • Li et al. [2018a] Tzu-Mao Li, Miika Aittala, Frédo Durand, and Jaakko Lehtinen. Differentiable monte carlo ray tracing through edge sampling. ACM Transactions on Graphics (TOG), 37(6):1–11, 2018a.
  • Li et al. [2017] Xiao Li, Yue Dong, Pieter Peers, and Xin Tong. Modeling surface appearance from a single photograph using self-augmented convolutional neural networks. ACM Transactions on Graphics (ToG), 36(4):1–11, 2017.
  • Li et al. [2018b] Zhengqin Li, Zexiang Xu, Ravi Ramamoorthi, Kalyan Sunkavalli, and Manmohan Chandraker. Learning to reconstruct shape and spatially-varying reflectance from a single image. ACM Transactions on Graphics (TOG), 37(6):1–11, 2018b.
  • Ma et al. [2007] Wan-Chun Ma, Tim Hawkins, Pieter Peers, Charles-Felix Chabert, Malte Weiss, Paul E Debevec, et al. Rapid acquisition of specular and diffuse normal maps from polarized spherical gradient illumination. Rendering Techniques, 2007(9):10, 2007.
  • Ma et al. [2023] Xiaohe Ma, Xianmin Xu, Leyao Zhang, Kun Zhou, and Hongzhi Wu. Opensvbrdf: A database of measured spatially-varying reflectance. ACM Trans. Graph, 42(6), 2023.
  • Mildenhall et al. [2020] Ben Mildenhall, Pratul P. Srinivasan, Matthew Tancik, Jonathan T. Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In ECCV, 2020.
  • Nam et al. [2018] Giljoo Nam, Joo Ho Lee, Diego Gutierrez, and Min H Kim. Practical svbrdf acquisition of 3d objects with unstructured flash photography. ACM Transactions on Graphics (TOG), 37(6):1–12, 2018.
  • Nocedal and Wright [1999] Jorge Nocedal and Stephen J Wright. Numerical optimization. Springer, 1999.
  • Rainer et al. [2020] Gilles Rainer, Abhijeet Ghosh, Wenzel Jakob, and Tim Weyrich. Unified neural encoding of btfs. In Computer Graphics Forum, pages 167–178. Wiley Online Library, 2020.
  • Solomon and Ikeuchi [1996] Fredric Solomon and Katsushi Ikeuchi. Extracting the shape and roughness of specular lobe objects using four light photometric stereo. IEEE Transactions on Pattern Analysis and Machine Intelligence, 18(4):449–454, 1996.
  • Srinivasan et al. [2021] Pratul P Srinivasan, Boyang Deng, Xiuming Zhang, Matthew Tancik, Ben Mildenhall, and Jonathan T Barron. Nerv: Neural reflectance and visibility fields for relighting and view synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7495–7504, 2021.
  • Tunwattanapong et al. [2013] Borom Tunwattanapong, Graham Fyffe, Paul Graham, Jay Busch, Xueming Yu, Abhijeet Ghosh, and Paul Debevec. Acquiring reflectance and shape from continuous spherical harmonic illumination. ACM Transactions on graphics (TOG), 32(4):1–12, 2013.
  • Ward [1992] Gregory J Ward. Measuring and modeling anisotropic reflection. In Proceedings of the 19th annual conference on Computer graphics and interactive techniques, pages 265–272, 1992.
  • Weinmann et al. [2014] Michael Weinmann, Juergen Gall, and Reinhard Klein. Material classification based on training data synthesized using a btf database. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part III 13, pages 156–171. Springer, 2014.
  • Woodham [1979] Robert J Woodham. Photometric stereo: A reflectance map technique for determining surface orientation from image intensity. In Image understanding systems and industrial applications I, pages 136–143. SPIE, 1979.
  • Yang et al. [2022] Jing Yang, Hanyuan Xiao, Wenbin Teng, Yunxuan Cai, and Yajie Zhao. Light sampling field and brdf representation for physically-based neural rendering. In The Eleventh International Conference on Learning Representations, 2022.
  • Ye et al. [2018] Wenjie Ye, Xiao Li, Yue Dong, Pieter Peers, and Xin Tong. Single image surface appearance modeling with self-augmented cnns and inexact supervision. In Computer Graphics Forum, pages 201–211. Wiley Online Library, 2018.
  • Zhang et al. [2019] Cheng Zhang, Lifan Wu, Changxi Zheng, Ioannis Gkioulekas, Ravi Ramamoorthi, and Shuang Zhao. A differential theory of radiative transfer. ACM Transactions on Graphics (TOG), 38(6):1–16, 2019.

A Proof

In this section, we provide detailed explanations for the equations discussed in the Preliminary and Method sections of the main paper.

Separating Diffuse and Specular Reflection.

The Fresnel equation reveals that specular reflection maintains the incident light’s polarization state, while diffuse reflection remains unpolarized. When the polarizer and analyzer are set perpendicular to each other (cross polarization), the analyzer blocks the specular reflection, allowing only the diffuse reflection to be measured as II_{\perp}. Conversely, when the polarizer and analyzer are aligned in parallel (parallel polarization), measured as II_{\parallel}, the specular reflection remains observable.

The state of polarization of light can be represented by a Stokes vector S=[S0,S1,S2,S3]TS=[S_{0},S_{1},S_{2},S_{3}]^{T} [9], where S0S_{0} is the total light intensity, S1S_{1} is the difference in intensity between horizontal and vertical linear polarization, S2S_{2} is the difference between linear polarization at π4\frac{\pi}{4} and π4-\frac{\pi}{4} and S3S_{3} is the difference between right-hand and left-hand circular polarization. The transformation of light’s polarization states via a linear polarizer at an angle θ\theta relative to a reference axis is defined by Mueller Matrics, M:

S=MSM=12[1cos2θsin2θ0cos2θcos22θcos2θsin2θ0sin2θcos2θsin2θsin22θ00000]\begin{split}S^{\prime}&=MS\\ M&=\frac{1}{2}\!\begin{bmatrix}1&\cos{2\theta}&\sin{2\theta}&0\\ \cos{2\theta}&\cos^{2}{2\theta}&\cos{2\theta}\sin{2\theta}&0\\ \sin{2\theta}&\cos{2\theta}\sin{2\theta}&\sin^{2}{2\theta}&0\\ 0&0&0&0\end{bmatrix}\end{split} (18)

When unpolarized light is incident on the polarizer, the input Stokes vector is S=[1,0,0,0]TS=[1,0,0,0]^{T}. As the light passes through a horizontally oriented polarizer, where θ=0\theta=0, the polarization state of light transforms from SS to SS^{\prime} defined as:

S=12[1cos0sin00cos0cos20cos0sin00sin0cos0sin0sin2000000][1000]=[1100]T\begin{split}S^{\prime}&=\frac{1}{2}\!\begin{bmatrix}1&\cos{0}&\sin{0}&0\\ \cos{0}&\cos^{2}{0}&\cos{0}\sin{0}&0\\ \sin{0}&\cos{0}\sin{0}&\sin^{2}{0}&0\\ 0&0&0&0\end{bmatrix}\begin{bmatrix}1\\ 0\\ 0\\ 0\end{bmatrix}\\ &=\begin{bmatrix}1&1&0&0\end{bmatrix}^{T}\end{split} (19)

Upon passing through the analyzer with an angle φ\varphi to the reference axis, the polarization state of the light further transforms from SS^{\prime} to S′′S^{\prime\prime} defined as:

S′′=12[1cos2φsin2φ0cos2φcos22φcos2φsin2φ0sin2φcos2φsin2φsin22φ00000][1100]=12[1+cos2φcos2φ+cos22φsin2φ+cos2φsin2φ0]=1+cos2φ2[1cos2φsin2φ0]=cos2φ[1cos2φsin2φ0]T\begin{split}S^{\prime\prime}&=\frac{1}{2}\!\begin{bmatrix}1&\cos{2\varphi}&\sin{2\varphi}&0\\ \cos{2\varphi}&\cos^{2}{2\varphi}&\cos{2\varphi}\sin{2\varphi}&0\\ \sin{2\varphi}&\cos{2\varphi}\sin{2\varphi}&\sin^{2}{2\varphi}&0\\ 0&0&0&0\end{bmatrix}\begin{bmatrix}1\\ 1\\ 0\\ 0\end{bmatrix}\\ &=\frac{1}{2}\begin{bmatrix}1+\cos{2\varphi}\\ \cos{2\varphi}+\cos^{2}{2\varphi}\\ \sin{2\varphi}+\cos{2\varphi}\sin{2\varphi}\\ 0\end{bmatrix}=\frac{1+\cos{2\varphi}}{2}\begin{bmatrix}1\\ \cos{2\varphi}\\ \sin{2\varphi}\\ 0\end{bmatrix}\\ &=\cos^{2}{\varphi}\begin{bmatrix}1&\cos{2\varphi}&\sin{2\varphi}&0\end{bmatrix}^{T}\end{split} (20)

Since the first number in S′′S^{\prime\prime} represents the total intensity, cos2φ\cos^{2}{\varphi} is the current intensity after the polarise and analyzer, which is also recognized as Malus’s Law (refere to Equation 1). Additionally, when adjusting the analyzer’s axis parallel (φ=0\varphi=0) or perpendicular (φ=π2\varphi=\frac{\pi}{2}) to the polarizer, the transmitted light can be simplified as follows:

S′′={[1100]Tif φ=00if φ=π2 S^{\prime\prime}=\begin{cases}\begin{bmatrix}1&1&0&0\end{bmatrix}^{T}&\text{if $\varphi=0$}\\ 0&\text{if $\varphi=\frac{\pi}{2}$ }\\ \end{cases} (21)

At φ=π2\varphi=\frac{\pi}{2}, the analyzer completely blocks the light. This characteristic can be leveraged to eliminate specular reflection. Therefore, the diffuse reflection can be represented via II_{\perp} while the specular reflection can be represented as the difference of two measurements represented via III_{\parallel}-I_{\perp}.

Moreover, on average, half of the light becomes polarized when passing through the polarizers [9], the total intensity from the diffuse reflection IdI_{d} and specular reflection IsI_{s} from the original unpolarized light can be derived via:

I=Id2π02πcos2θdθ=Id12II=Is2π02πcos2θdθId=2IIs=2I2I\begin{split}I_{\perp}&=\frac{I_{d}}{2\pi}\int_{0}^{2\pi}\cos^{2}{\theta}d\theta=I_{d}\cdot\frac{1}{2}\\ I_{\parallel}-I_{\perp}&=\frac{I_{s}}{2\pi}\int_{0}^{2\pi}\cos^{2}{\theta}d\theta\\ I_{d}&=2I_{\perp}\\ I_{s}&=2I_{\parallel}-2I_{\perp}\end{split} (22)

Initial Albedo Estimation.

An initial estimation of diffuse and specular albedo ρdinit\rho_{d}^{\text{init}}, ρsinit\rho_{s}^{\text{init}} can be roughly derived from the separated diffuse and specular reflection IdI_{d}, IsI_{s}. This estimation relies on the rendering equation for light transport at each surface point that inherently considers albedo.

In the presence of uniform white illumination, the observed radiant intensities for both diffuse reflection IdI_{d} and specular reflection IsI_{s} at any given surface point pp can be described as follows:

Id,p=ωiΩρd,pL(ωi)(npωi)𝑑ωiI_{d,p}=\int_{\omega_{i}\in\Omega}\rho_{d,p}\cdot L(\omega_{i})\cdot(n_{p}\cdot\omega_{i})\,d\omega_{i} (23)
Is,p=ωiΩρs,pL(ωi)fσ(ωi,ωo)𝑑ωiI_{s,p}=\int_{\omega_{i}\in\Omega}\rho_{s,p}\cdot L(\omega_{i})\cdot f_{\sigma}(\omega_{i},\omega_{o})\,d\omega_{i} (24)

, where ωi\omega_{i} and ωo\omega_{o} represent the incoming and outgoing lighting directions, respectively, with L(ωi)L(\omega_{i}) denoting the incident radiance from the ωi\omega_{i}, which remains uniform over the sphere. Additionally, ρd,p\rho_{d,p} and ρs,p\rho_{s,p} correspond to the diffuse and specular albedo at point pp, while npn_{p} represents the surface point normal, and fσf_{\sigma} stands for the specular reflection distribution function.

We can establish a local frame where the Y-axis aligns with the surface normal nn. Therefore, we can rewrite the ωi\omega_{i} in spherical coordinates as ωi=[sin(θ)cos(ϕ),sin(θ)sin(ϕ),cos(θ)]\omega_{i}=[\sin(\theta)\cos(\phi),\sin(\theta)\sin(\phi),\cos(\theta)], with θ\theta as the polar angle and ϕ\phi as the azimuthal angle within the local frame. Moreover, under uniform lighting conditions, the incident radiance L(ωi)L(\omega_{i}) over the spherical sphere is a constant L0L_{0}. This allows us to further derive ρd\rho_{d} and ρs\rho_{s} over the entire observation as follows:

ρd=Id(ωiΩL(ωi)(nωi)𝑑ωi)1=Id(L0ωiΩ(nωi)𝑑ωi)1=Id(L002π0π/2cos(θ)sin(θ)𝑑θ𝑑ϕ)1=Id(L002π12𝑑ϕ)1=Id(L0π)1\begin{split}\rho_{d}&=I_{d}\cdot\left(\int_{\omega_{i}\in\Omega}L(\omega_{i})\cdot(n\cdot\omega_{i})\,d\omega_{i}\right)^{-1}\\ &=I_{d}\cdot\left(L_{0}\cdot\int_{\omega_{i}\in\Omega}(n\cdot\omega_{i})\,d\omega_{i}\right)^{-1}\\ &=I_{d}\cdot\left(L_{0}\cdot\int_{0}^{2\pi}\int_{0}^{\pi/2}\cos(\theta)\sin(\theta)\,d\theta d\phi\right)^{-1}\\ &=I_{d}\cdot\left(L_{0}\cdot\int_{0}^{2\pi}\frac{1}{2}d\phi\,\right)^{-1}\\ &=I_{d}\cdot(L_{0}\cdot\pi)^{-1}\\ \end{split} (25)
ρs=Is(ωiΩL(ωi)fσ(ωi,ωo)𝑑ωi)1=Is(L0ωiΩfσ(ωi,ωo)𝑑ωi)1=Is(L01)1\begin{split}\rho_{s}&=I_{s}\cdot\left(\int_{\omega_{i}\in\Omega}L(\omega_{i})\cdot f_{\sigma}(\omega_{i},\omega_{o})\,d\omega_{i}\right)^{-1}\\ &=I_{s}\cdot\left(L_{0}\cdot\int_{\omega_{i}\in\Omega}f_{\sigma}(\omega_{i},\omega_{o})\,d\omega_{i}\right)^{-1}\\ &=I_{s}\cdot\left(L_{0}\cdot 1\right)^{-1}\end{split} (26)

Also, the ωiΩ(nωi)𝑑ωi\int_{\omega_{i}\in\Omega}(n\cdot\omega_{i})\,d\omega_{i} can also be simulated in the local frame via the law of large numbers:

y=(2π)1ωiΩ(nωi)𝑑ωi=limm1mk=0mmax(nωik,0)=21\begin{split}y&=(2\pi)^{-1}\cdot\int_{\omega_{i}\in\Omega}(n\cdot\omega_{i})\,d\omega_{i}\\ &=\lim_{m\rightarrow\infty}\frac{1}{m}\sum_{k=0}^{m}\max(n\cdot\omega_{i}^{k},0)\\ &=2^{-1}\\ \end{split} (27)

By conducting uniform sampling of ωi\omega_{i} from a unit sphere, we simulate the results as shown in Table 2.

m 10310^{3} 10410^{4} 10510^{5} 10610^{6} 10710^{7} 10810^{8}
y 0.519 0.503 0.499 0.499 0.500 0.500
Table 2: Law of Large Numbers over Equation 27. mm is the number of samples over the unit sphere and yy is the results.

Moreover, given that Ward’s model [44] subject to 2D normal distribution, denoted as fσ(ωi,ωo)𝒩(μ,σ2)f_{\sigma}(\omega_{i},\omega_{o})\sim\mathcal{N}(\mu,\sigma^{2}). It’s expected that the overall integral of fσ(ωi,ωo)f_{\sigma}(\omega_{i},\omega_{o}) over Ω\Omega equals 1. Further details can be found in [19].

During the capture, measurements become discrete through OLAT, with each light covering a specific area denoted as A0A_{0} over the entire spherical surface. Consequently, we approximate the results via the captured sequence Λ={Idi}i=0N\Lambda=\{I_{d}^{i}\}_{i=0}^{N}:

ρdinit=Id(ωiΩL(ωi)(nωi)𝑑ωi)1=Id(L0π)1=2πωiΩRd(ωi,ωo)𝑑ωi2π(L0π)1=limm(2πmIdim21A0)(L0π)12πi=0NIdiN21A0(L0π)1=(4πNA0i=0NIdi)(L0π)1=4NL0A0i=0NIdi\begin{split}\rho_{d}^{\text{init}}&=I_{d}\cdot\left(\int_{\omega_{i}\in\Omega}L(\omega_{i})\cdot(n\cdot\omega_{i})\,d\omega_{i}\right)^{-1}\\ &=I_{d}\cdot(L_{0}\cdot\pi)^{-1}\\ &=2\pi\frac{\int_{\omega_{i}\in\Omega}R_{d}(\omega_{i},\omega_{o})d\omega_{i}}{2\pi}\cdot(L_{0}\cdot\pi)^{-1}\\ &=\lim_{m\rightarrow\infty}\left(2\pi\frac{\sum_{m}I_{d}^{i}}{m\cdot 2^{-1}\cdot A_{0}}\right)\cdot(L_{0}\cdot\pi)^{-1}\\ &\approx 2\pi\frac{\sum_{i=0}^{N}I_{d}^{i}}{N\cdot 2^{-1}\cdot A_{0}}\cdot(L_{0}\cdot\pi)^{-1}\\ &=\left(\frac{4\pi}{N\cdot A_{0}}\sum_{i=0}^{N}I_{d}^{i}\right)\cdot(L_{0}\cdot\pi)^{-1}\\ &=\frac{4}{N\cdot L_{0}\cdot A_{0}}\sum_{i=0}^{N}I_{d}^{i}\\ \end{split} (28)
ρsinit=Is(ωiΩL(ωi)fσ(ωi,ωo)𝑑ωi)1=Is(L01)1=2πωiΩRs(ωi,ωo)𝑑ωi2π(L01)1=limm(2πmIsim21A0)(L01)1(4πNA0i=0NIsi)(L01)1=4πNL0A0i=0NIsi\begin{split}\rho_{s}^{\text{init}}&=I_{s}\cdot\left(\int_{\omega_{i}\in\Omega}L(\omega_{i})\cdot f_{\sigma}(\omega_{i},\omega_{o})\,d\omega_{i}\right)^{-1}\\ &=I_{s}\cdot\left(L_{0}\cdot 1\right)^{-1}\\ &=2\pi\frac{\int_{\omega_{i}\in\Omega}R_{s}(\omega_{i},\omega_{o})d\omega_{i}}{2\pi}\cdot(L_{0}\cdot 1)^{-1}\\ &=\lim_{m\rightarrow\infty}\left(2\pi\frac{\sum_{m}I_{s}^{i}}{m\cdot 2^{-1}\cdot A_{0}}\right)\cdot\left(L_{0}\cdot 1\right)^{-1}\\ &\approx\left(\frac{4\pi}{N\cdot A_{0}}\sum_{i=0}^{N}I_{s}^{i}\right)\cdot(L_{0}\cdot 1)^{-1}\\ &=\frac{4\pi}{N\cdot L_{0}\cdot A_{0}}\sum_{i=0}^{N}I_{s}^{i}\\ \end{split} (29)

where RdR_{d} and RsR_{s} denote the diffuse reflection function and the specular reflection function of the material, respectively. NN is the number of lighting directions. For further simplification, we introduce the constant κ=(L0A0)1\kappa=(L_{0}\cdot A_{0})^{-1}. As a result, the equation can be expressed as:

ρdinit=4κNi=0NIdi\rho_{d}^{\text{init}}=\frac{4\kappa}{N}\sum\nolimits_{i=0}^{N}I_{d}^{i} (30)
ρsinit=4πκNi=0NIsi\rho_{s}^{\text{init}}=\frac{4\pi\kappa}{N}\sum\nolimits_{i=0}^{N}I_{s}^{i} (31)

Occlusion τ\tau

Occlusion describes the overall visibility ν\nu of incident lighting ωi\omega_{i} from the upper hemisphere Ω\Omega when observed via ωo\omega_{o}. This is mathematically expressed as 12πΩν(ωi,ωo)(ωin)𝑑ωi\frac{1}{2\pi}\int_{\Omega}\nu(\omega_{i},\omega_{o})\cdot(\omega_{i}\cdot n)d\omega_{i}. Notably, since the average of ωin\omega_{i}\cdot n is 12\frac{1}{2} over the hemisphere, the occlusion remains at 12\frac{1}{2} when incident lighting ωi\omega_{i} from any solid angle is visible via ωo\omega_{o}. To get the normalized occlusion over the hemisphere, we factor out the average of ωin\omega_{i}\cdot n, resulting in:

τ=12πΩν(ωi,ωo)(ωin)𝑑ωi(12πΩ(ωin)𝑑ωi)1=12πlimm(2πmIim21)(12)14Nk=0NIkζmax(ωikn,0)\begin{split}\tau&=\frac{1}{2\pi}\int_{\Omega}\nu(\omega_{i},\omega_{o})\cdot(\omega_{i}\cdot n)d\omega_{i}\cdot\left(\frac{1}{2\pi}\int_{\Omega}(\omega_{i}\cdot n)d\omega_{i}\right)^{-1}\\ &=\frac{1}{2\pi}\lim_{m\rightarrow\infty}\left(2\pi\frac{\sum_{m}I^{i}}{m\cdot 2^{-1}}\right)\cdot(\frac{1}{2})^{-1}\\ &\approx\frac{4}{N}\sum_{k=0}^{N}\lceil I^{k}-\zeta\rceil\cdot max(\omega_{i}^{k}\cdot n,0)\end{split} (32)

where ζ\zeta is the average ambient noise map and \lceil\cdot\rceil is the ceiling operator.

B Implementation Details and Limitations

Refer to caption
Figure 12: Dataset Examples.

Storage and Memory Cost

Our capture process takes place in a Light Stage with 8 RED KOMODO 6K cameras, synchronized at 30 frames per second. For each scan, the captured data, saved as 3240×61443240\times 6144 resolution R3D files. Prior to using the data from the raw scan, we extract individual frames stored in OpenEXR format. In this work, we conducted complete Polarized OLAT scans for 26 objects. We present frontal views of several objects in our dataset in Figure 12 and further present the polarized OLAT captures for a specific object from multiview in Figure 20.

GPU Usage

Our implementation leverages JAX [7] for efficient GPU access, ensuring a lightweight solution. We also employ JAXopt [4] for optimization, enabling batchable and differentiable solvers on large data blocks with a single workstation. We perform batch processing via vectorizing map on an RTX A6000, employing a batch size of 2202^{20} for optimizing diffuse reflection and 2182^{18} for optimizing specular reflection.

Lens Flare and Inter-reflection Constrains

We implement the constraints n×ωi0n\times\omega_{i}\geq 0 to mitigate the impact of lens flare and inter-reflection. In practical terms, we apply these constraints to filter out values that do not meet the criteria before optimization. This approach reduces computational complexity and facilitates the optimizer’s convergence to the optimal solution. Unconstrained data typically introduces a considerable number of zeros, which usually leads the optimizer to generate blank results. The constraints effectively address this issue.

Noise Pixels

Throughout the optimization process, the presence of noise pixels (usually in the background) can significantly prolong the solver’s search for the local minimum, and unfortunately, this extended search doesn’t lead to a meaningful solution. To address this challenge, we choose to terminate the solver when the linear search encounters failure, which is often a consequence of noise pixels in the background. Notably, the solver tends to converge more readily when dealing with pixels located in the foreground objects.

Limitations

The proposed method successfully decomposes highly glossy materials. However, capturing living creatures or humans presents challenges due to the current setup’s requirement to capture a dense reflectance field both with and without polarization. This process typically takes around 10 seconds given the current frame rate, and any slight movement of the subject can lead to color bleeding issues. Future implementations will need to incorporate frame tracking to address these movements effectively. Additionally, the method does not currently measure geometry, leading to difficulties in managing shadows caused by self-occlusion. This could be mitigated by integrating geometry data from multiview captures.

C More Experiments and Results

C.1 Ablation Studies

Runtime \downarrow n^d\hat{n}_{d} ρ^d\hat{\rho}_{d} n^s\hat{n}_{s} ρ^s\hat{\rho}_{s} σ^\hat{\sigma}
LBFGS (backtracking) 1.54e-6 1.63e-6 1.75e-5 8.21e-6 2.06e-5
LBFGS (zoom) 2.98e-6 3.37e-6 4.45e-5 1.43e-5 2.20e-5
LBFGS (hager-zhang) 7.94e-6 4.81e-6 5.79e-5 2.83e-5 9.38e-5
GD 3.01e-6 2.23e-6 2.33e-5 4.31e-6 4.56e-6
NCG 3.39e-5 4.42e-6 4.77e-4 6.75e-5 6.78e-5
GN - 8.33e-7 - 6.40e-6 5.97e-5
Error \downarrow n^d\hat{n}_{d} ρ^d\hat{\rho}_{d} n^s\hat{n}_{s} ρ^s\hat{\rho}_{s} σ^\hat{\sigma}
LBFGS (backtracking) 3.60e-4 3.73 2.02e-3 3.94 7.31e-1
LBFGS (zoom) 7.78e-5 5.31e-5 4.34e-4 7.87e-5 3.17e-1
LBFGS (hager-zhang) 7.64e-5 4.85e-8 3.29e-4 2.49e-6 5.24e-2
GD 6.23e-2 5.12e-4 2.54e-1 5.13e-3 5.53e-1
NCG 2.20e-4 2.74e-5 8.72e-4 1.13e-4 1.28e-3
GN - 5.18e-10 - 1.16e-8 2.21e-3
Table 3: Solver Runtime (seconds) and Errors. Lower is better (\downarrow). We emphasize the max and min in the column accordingly.

Ablation Study on Optimization

Further ablation results are presented in Figure 16 and 17 to affirm the quality enhancements achieved through optimization on both diffuse and specular normals. In the absence of optimization, the obtained normal maps may contain baked-in color artifacts, making the normal distribution sensitive to surface color variations. However, this issue is mitigated upon the introduction of optimization. Additionally, specular normals may exhibit blending with overexposure values, leading to noise in the data. The optimization process efficiently reduces such noise, contributing to overall improvement in quality.

Refer to caption
Figure 13: Ablation study on overexposure removal: Specular Albedo. The proposed overexposure removal effectively mitigates the lighting baked-in effect from the acquired specular albedo
Refer to caption
Figure 14: Ablation Study on overexposure removal: Diffuse Albedo. Similar to the improvements over specular albedo, overexposure removal can also improve diffuse albedo.
Refer to caption
Figure 15: Ablation Study on Overexposure Removal: Iterations We showcase the diffuse albedo and specular albedo obtained with overexposure removal via various iterations in (b, d), and the corresponding removed overexposure values in (a, c), where M is the total number of iterations in the Overexposure Removal Algorithm.
Refer to caption
Figure 16: Ablation Study on acquired diffuse normal with and without optimization.
Refer to caption
Figure 17: Ablation Study on acquired specular normal with and without optimization.

Ablation Study on Overexposure Removal

Additional ablation results are presented in Figure 13 and 14, illustrating the impact of overexposure removal on both specular albedo ρs\rho_{s} and diffuse albedo ρd\rho_{d}. Particularly for objects with pronounced specular surfaces, our overexposure removal method effectively eliminates artifacts from incoming light sources while preserving intricate surface details. This efficacy is further demonstrated through visualizations showcasing different values of M, representing total iterations, in the Overexposure Removal Algorithm (see Figure 15).

The algorithm treats intensity variations as a sequential signal and addresses anomalies accordingly. Typically, during scanning, the maximum intensity values in the signal result from overexposure and offer limited useful reflection information. In practice, we set M=2 to efficiently remove overexposure while retaining the original intensity distribution to the maximum extent possible. This choice strikes a balance between removing overexposure artifacts and preserving valuable reflection data.

Optimization Solvers

We compare the results obtained from various optimization solvers, including LBFGS (backtracking [3]), LBFGS (zoom [39]), LBFGS (hager-zhang [24]), Gradient Descent (GD), and Nonlinear Conjugate Gradient (NCG [3]), for solving n^d\hat{n}_{d}, ρ^d\hat{\rho}_{d}, n^s\hat{n}_{s}, and ρ^s\hat{\rho}_{s}. Additionally, we evaluate the performance of the Gauss-Newton (GN) nonlinear optimization approach for solving ρ^d\hat{\rho}_{d}, ρ^s\hat{\rho}_{s}, and σ^\hat{\sigma}. In this context, optimizing surface normals n^d\hat{n}_{d} and n^s\hat{n}_{s} is inappropriate, as the cost function relies on correlation. The results are tested on RTX A6000 and averaged per pixel as shown in Table 3. Note that errors are measured as the L2-norm of the gradient vector upon solver convergence or reaching the maximum iterations, 500 in our case.

LBFGS (backtracking) achieves fast convergence but with higher errors, especially in albedo optimization, whereas LBFGS (zoom) and LBFGS (hager-zhang) require more time but offer improved accuracy. To balance runtime and error, we use LBFGS (backtracking) for normal optimization, LBFGS (zoom) for albedo optimization, and Gauss-Newton for σ\sigma optimization. More implementation details can be found in the supplemental material.

C.2 More Comparisons

In Figure 18 and 19, we provide more qualitative comparisons between our results and [35], using static gradient illumination capture. Traditional methods encounter challenges when it comes to generating clean albedo for specular objects. Additionally, these methods struggle to effectively remove the undesired texture patterns from the measured normal maps. In contrast, our proposed method consistently surpasses traditional approaches, resulting in overall improved outcomes.

C.3 Other Results

More Optimization Results

As depicted in Figure 21, this section provides additional results following our material optimization from multiple viewpoints. These results encompass a diverse range of shapes, from standard geometric forms to everyday objects, as well as materials spanning from relatively diffuse to highly specular. The materials demonstrate view consistency in the optimized maps and exhibit robustness against rotations.

Refer to caption
Figure 18: More Qualitative Comparison. We compare results from 1) our method with 2) from [35] via static capture on objects with specular outer layers. Examined properties cover diffuse albedo ρd\rho_{d} and specular albedo ρs\rho_{s}, diffuse normal ndn_{d}, and specular normal nsn_{s}, with zoomed-in views. Normally, diffuse and specular normals are similar, but in multi-layered materials, they may differ slightly.
Refer to caption
Figure 19: More Qualitative Comparison. We compare results from 1) our method with 2) from [35] via static capture on objects with specular outer layers. Examined properties cover diffuse albedo ρd\rho_{d} and specular albedo ρs\rho_{s}, diffuse normal ndn_{d}, and specular normal nsn_{s}, with zoomed-in views. Normally, diffuse and specular normals are similar, but in multi-layered materials, they may differ slightly.
Refer to caption
Figure 20: Polarized OLAT from Multiviews. We showcase an example captured object from multiview under cross-polarized and parallel-polarized OLAT.
Refer to caption
Figure 21: More Optimization Results. We present a grey ball from a) right and b) front, a soda can from c) left and d) right, and a specular cup from e) right and f) front. For each object, we showcase 1) original image, 2) diffuse albedo ρ^d\hat{\rho}_{d}, 3) specular albedo ρ^s\hat{\rho}_{s}, 4) diffuse normal n^d\hat{n}_{d}, 5) specular normal n^s\hat{n}_{s}, 6) anisotropy ϱ\varrho, and 7) roughness γ\gamma.