Pixel2Mesh++: Multi-View 3D Mesh Generation via Deformation


Fudan University     Google LLC     Nuro, Inc.


Pixel2Mesh++ Shape Generation Pipeline.



Abstract

We study the problem of shape generation in 3D mesh representation from a few color images with known camera poses. While many previous works learn to hallucinate the shape directly from priors, we resort to further improving the shape quality by leveraging cross-view information with a graph convolutional network. Instead of building a direct mapping function from images to 3D shape, our model learns to predict series of deformations to improve a coarse shape iteratively. Inspired by traditional multiple view geometry methods, our network samples nearby area around the initial mesh's vertex locations and reasons an optimal deformation using perceptual feature statistics built from multiple input images. Extensive experiments show that our model produces accurate 3D shape that are not only visually plausible from the input perspectives, but also well aligned to arbitrary viewpoints. With the help of physically driven architecture, our model also exhibits generalization capability across different semantic categories, number of input images, and quality of mesh initialization.




Network Architecture




Paper, Code and Data

C. Wen, Y. Zhang, Z. Li, Y. Fu.

Pixel2Mesh++: Multi-View 3D Mesh Generation via Deformation.

ICCV, 2019.

[arXiv]     [Bibtex]     [GitHub]     [Google Drive]



Results



Results on ShapeNet objects.
Robustness to Initialization.



Acknowledgements

This work is supported by the STCSM project (19ZR1471800), and Eastern Scholar (TP2017006). We would also like to thank Pixel2Mesh. The websiteis modified from this template.