Neural Point-based Shape Modeling of Humans in Challenging Clothing

In 3DV 2022


Qianli Ma 1,2 Jinlong Yang 2 Michael J. Black 2 Siyu Tang 1

1 ETH Zürich
2 Max Planck Institute for Intelligent Systems, Tübingen



Paper

Code

Video

TL;DR

Complex clothing types such as skirts and dresses are still challenging today even for SoTA learning-based digital human models (see below image). We address this challenge with SkiRT: our two-stage point-based representation that learns clothing templates, skinning weight fields and pose-dependent deformation fields from 3D scans.


Teaser figure.


Abstract

Parametric 3D body models like SMPL only represent minimally-clothed people and are hard to extend to clothing because they have a fixed mesh topology and resolution. To address this limitation, recent work uses implicit surfaces or point clouds to model clothed bodies. While not limited by topology, such methods still struggle to model clothing that deviates significantly from the body, such as skirts and dresses. This is because they rely on the body to canonicalize the clothed surface by reposing it to a reference shape. Unfortunately, this process is poorly defined when clothing is far from the body. Additionally, they use linear blend skinning to pose the body and the skinning weights are tied to the underlying body parts. In contrast, we model the clothing deformation in a local coordinate space without canonicalization. We also relax the skinning weights to let multiple body parts influence the surface. Specifically, we extend point-based methods with a coarse stage, that replaces canonicalization with a learned pose-independent “coarse shape” that can capture the rough surface geometry of clothing like skirts. We then refine this using a network that infers the linear blend skinning weights and pose dependent displacements from the coarse representation. The approach works well for garments that both conform to, and deviate from, the body. We demonstrate the usefulness of our approach by learning person-specific avatars from examples and then show how they can be animated in new poses and motions. We also show that the method can learn directly from raw scans with missing data, greatly simplifying the process of creating realistic avatars. Code is available for research purposes.




Paper


Neural Point-based Shape Modeling of Humans in Challenging Clothing
Qianli Ma, Jinlong Yang, Michael J. Black and Siyu Tang.
In 3DV 2022
[Paper]  [Supp]  [arXiv]
@inproceedings{SkiRT:3DV:2022,
title = {Neural Point-based Shape Modeling of Humans in Challenging Clothing},
author = {Ma, Qianli and Yang, Jinlong and Black, Michael J. and Tang, Siyu},
booktitle = {2022 International Conference on 3D Vision (3DV)},
month = September,
year = {2022},
month_numeric = {9}}



Related Projects

The Power of Points for Modeling Humans in Clothing (ICCV 2021)
Qianli Ma, Jinlong Yang, Siyu Tang, Michael J. Black
The former SoTA on ponit-based digital human models, PoP: a point-based, unified model for multiple subjects and outfits that can turn a single, static 3D scan into an animatable avatar with natural pose-dependent clothing deformations. In this paper we also introduced ReSynth: a high-quality synthetic dataset of challenging clothing includig a variety of skirts and dresses.

SCALE: Modeling Clothed Humans with a Surface Codec of Articulated Local Elements (CVPR 2021)
Qianli Ma, Shunsuke Saito, Jinlong Yang, Siyu Tang, Michael J. Black
Our first point-based model for humans: modeling pose-dependent shapes of clothed humans explicitly with hundreds of articulated surface elements: the clothing deforms naturally even in the presence of topological change!




Acknowledgements

We thank Shaofei Wang for insightful discussions. Qianli Ma acknowledges the support from the Max Planck ETH Center for Learning Systems. This webpage is adapted from the template of Neural Parts.