论文标题
多视纹理网状恢复通过可区分渲染
Multiview Textured Mesh Recovery by Differentiable Rendering
论文作者
论文摘要
尽管通过自学实现了形状和颜色恢复的有希望的结果,但基于多层感知的方法通常会在学习深层隐式表面表示方面遭受沉重的计算成本。由于渲染每个像素需要一个前向网络推断,因此综合整个图像是非常密集的。为了应对这些挑战,我们提出了一种有效的粗到精细方法,以从本文中从多视图中恢复纹理网格。具体而言,采用了可区分的泊松求解器来表示对象的形状,该形状能够产生拓扑 - 敏捷和水密表面。为了说明深度信息,我们通过最小化渲染网格与多视图立体声的预测深度之间的差异来优化形状几何形状。与隐式神经表示形状和颜色相反,我们引入了一种基于物理的逆渲染方案,以共同估计环境照明和对象的反射率,这能够实时呈现高分辨率图像。重建的网格的质地是从可学习的密集纹理网格中插值的。我们已经对几个多视图立体数据集进行了广泛的实验,其有前途的结果证明了我们提出的方法的功效。该代码可在https://github.com/l1346792580123/diff上找到。
Although having achieved the promising results on shape and color recovery through self-supervision, the multi-layer perceptrons-based methods usually suffer from heavy computational cost on learning the deep implicit surface representation. Since rendering each pixel requires a forward network inference, it is very computational intensive to synthesize a whole image. To tackle these challenges, we propose an effective coarse-to-fine approach to recover the textured mesh from multi-views in this paper. Specifically, a differentiable Poisson Solver is employed to represent the object's shape, which is able to produce topology-agnostic and watertight surfaces. To account for depth information, we optimize the shape geometry by minimizing the differences between the rendered mesh and the predicted depth from multi-view stereo. In contrast to the implicit neural representation on shape and color, we introduce a physically based inverse rendering scheme to jointly estimate the environment lighting and object's reflectance, which is able to render the high resolution image at real-time. The texture of the reconstructed mesh is interpolated from a learnable dense texture grid. We have conducted the extensive experiments on several multi-view stereo datasets, whose promising results demonstrate the efficacy of our proposed approach. The code is available at https://github.com/l1346792580123/diff.