Specular-to-Diffuse Translation for Multi-View Reconstruction

Shihao Wu, Hui Huang, Tiziano Portenier, Matan Sela, Daniel Cohen-Or, Ron Kimmel, Matthias Zwicker; The European Conference on Computer Vision (ECCV), 2018, pp. 183-200


Most multi-view 3D reconstruction algorithms, especially when shape-from-shading cues are used, assume that object appearance is predominantly diffuse. To alleviate this restriction, we introduce S2Dnet, a generative adversarial network for transferring multiple views of objects with specular reflection into diffuse ones, so that multi-view reconstruction methods can be applied more effectively. Our network extends unsupervised image-to-image translation to multi-view ``specular to diffuse" translation. To preserve object appearance across multiple views, we introduce a Multi-View Coherence loss (MVC) that evaluates the similarity and faithfulness of local patches after the view-transformation. Our MVC loss ensures that the similarity of local correspondences among multi-view images is preserved under the image-to-image translation. As a result, our network yields significantly better results than several single-view baseline techniques. In addition, we carefully design and generate a large synthetic training data set using physically-based rendering. During testing, our network takes only the raw glossy images as input, without extra information such as segmentation masks or lighting estimation. Results demonstrate that multi-view reconstruction can be significantly improved using the images filtered by our network. We also show promising performance on real world training and testing data.

Related Material

author = {Wu, Shihao and Huang, Hui and Portenier, Tiziano and Sela, Matan and Cohen-Or, Daniel and Kimmel, Ron and Zwicker, Matthias},
title = {Specular-to-Diffuse Translation for Multi-View Reconstruction},
booktitle = {The European Conference on Computer Vision (ECCV)},
month = {September},
year = {2018}