You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Our method focuses on generating high-resolution shading estimations, rather than attempting to estimate shading and albedo separately.
15
15
Since shading values are unbounded, we develop a representation of shading values called "inverse shading" which maps the shading values into the zero-one range.
16
16
This creates a balanced distribution of values in a well-defined range that is desirable for training neural networks.
Rather than directly regressing the inverse shading values, we relax the problem and aim to predict *ordinal* shading values.
21
21
To do this, we train our network using shift- and scale-invariant loss functions.
22
22
This simplifies the task of shading estimation as the model does not need to estimate precise values that satisfy the core intrinsic decomposition model
To generate a final shading estimation we combine two ordinal estimations, at low and high resolutions, with the input image and send them through a second network.
33
33
We use the final shading estimation, and the input image in order to compute our estimated albedo. This allows us to compute losses on both shading and albedo while
We train our method on multiple rendered datasets. In order to generate real-world supervision for our method we use multi-illumination data.
39
39
Using our pipeline we estimate the albedo for each image in a given multi-illumination scene. By taking the median across these albedo estimations, small errors are removed resulting in a single accurate albedo.
40
40
We use these 25,000 pseudo-ground-truth pairs as training data and continue training our pipeline.
0 commit comments