Unsupervised Shape and Pose Disentanglement for 3D Meshes

Keyang Zou, Bharat Lal Bhatnagar, Gerard Pons-Moll

Max Planck Institute for Informatics, Saarland Informatics Campus, Germany
Paper, Supplementary, Code, Video, Arxiv

Abstract

Parametric models of humans, faces, hands and animals have been widely used for a range of tasks such as image-based reconstruction, estimating correspondence between 3D shapes, or animation. Their key strength is the ability to factor surface variations into shape and pose dependent components. The process of learning such models requires lots of expert knowledge and hand-defined object specific constraints, making the learning approach unscalable to novel object categories. In this paper, we present a simple yet effective approach to learn disentangled shape and pose representations in an unsupervised setting. We use a combination of self-consistency and cross-consistency constraints to learn pose and shape space from a dataset of registered meshes. We additionally incorporate as-rigid-as-possible surface deformation(ARAP) into the training loop to avoid degenerate solutions. We demonstrate the usefulness of learned representations through a number of tasks including pose transfer and shape retrieval. The experiments on datasets of 3D humans, faces, hands and animals demonstrate the generality of our approach.

Pose Transfer

Pose & Shape Interpolation

Code

Citation

@InProceedings{keyang20unsupervised,
      author = {Zhou, Keyang and Bhatnagar, Bharat Lal and Pons-Moll, Gerard},
      title = {Unsupervised Shape and Pose Disentanglement for 3D Meshes},
      booktitle = {The European Conference on Computer Vision (ECCV)},
      month = {August},
      year = {2020},
  }