3D Pose Transfer with Correspondence Learning and Mesh Refinement
NeurIPS 2021
Chaoyue Song
Jiacheng Wei
Ruibo Li
Fayao Liu
Guosheng Lin
[Paper]
[GitHub]
[Video]

Abstract

3D pose transfer is one of the most challenging 3D generation tasks. It aims to transfer the pose of a source mesh to a target mesh and keep the identity (e.g., body shape) of the target mesh. Some previous works require key point annotations to build reliable correspondence between the source and target meshes, while other methods do not consider any shape correspondence between sources and targets, which leads to limited generation quality. In this work, we propose a correspondence-refinement network to help the 3D pose transfer for both human and animal meshes. The correspondence between source and target meshes is first established by solving an optimal transport problem. Then, we warp the source mesh according to the dense correspondence and obtain a coarse warped mesh. The warped mesh will be better refined with our proposed Elastic Instance Normalization, which is a conditional normalization layer and can help to generate highquality meshes. Extensive experimental results show that the proposed architecture can effectively transfer the poses from source to target meshes and produce better results with satisfied visual performance than state-of-the-art methods.


Poster



Result video



Method

We solve the pose transfer problem with our proposed correspondence-refinement network. We learn the shape correspondence by solving an optimal transport problem without any key point annotations and generate high-quality final meshes with our proposed elastic instance normalization in the refinement module.



Results on human and animal meshes



Comparison with other methods on human meshes

The identity and pose meshes are from SMPL. Our method and DT (needs key point annotations) can generate better results than Wang et al. when doing pose transfer on human meshes. The results generated by Wang et al. are always not smooth on the arms or legs. Since DT needs user to label the key point annotations, our method is more efficient and practical than DT.



Comparison with other methods on animal meshes

The identity and pose meshes are from SMAL. Our method produces more successful results when doing pose transfer on different animal meshes. Although DT has key point annotations, it still fails to transfer the pose when the identity of the mesh pairs are very different. The method of Wang et al. produces very flat legs and wrong direction faces.



Generalization capability

To evaluate the generalization capability of our method, we evaluate it on FAUST and MG-dataset. Human meshes in FAUST have the same number of vertices as SMPL and have more unseen identities. In MG-dataset, the human meshes are all dressed which have 27554 vertices each and have more realistic details. Our method can also have a good performance on FAUST and MG-dataset.



Paper

C. Song, J. Wei, R. Li, F. Liu, G. Lin.
3D Pose Transfer with Correspondence Learning and Mesh Refinement.
NeurIPS, 2021.
(hosted on ArXiv)


[Bibtex]


Acknowledgements

This template was originally made by Phillip Isola and Richard Zhang for a colorful ECCV project; the code can be found here.