Skip to content

MengZephyr/Motion-Guided-Deep-Dynamic-3D-Garment

Repository files navigation

Motion-Guided-Deep-Dynamic-3D-Garment

Introduction

This repository contains the implemetation of Motion-Guided-Deep-Dynamic-3D-Garment.
Our approach first learns a compact generative space of plausible garment deformations. We achieve this by encoding a garment geometry represented as relative to the underlying body to a latent code using a feature map encoder. A decoder then predicts a geometry feature map. We sample the feature map to obtain per-vertex geometry features. These features along with the vertex UV coordinates are provided to an MLP, to predict per-vertex canonical space displacements. We assign each vertex a skinning weight based on its proximity to the underlying body seed points weighted by a per-body part learnable kernel radius. Once a generative space of garment deformations are learned, we train a dynamic-aware encoder. We provide the previous garment geometry, the garment velocity, acceleration and the interaction between the body and the garment as input. The dynamic-aware encoder maps these inputs to a latent code in the learned deformation space which is then used to decode the current garment geometry. Blocks denoted in blue are pre-trained and kept fixed when training the blocks in green. More details, please refer to our paper.

Results


We train our network on a walking sequence of 300 frames on a fixed body shape and test on walking motion with different character armspace settings, different styles of walking, different body shapes, and even challenging dancing motions.

Have fun!

If you want to setup your own training, you need to prepare the dataset composed of the mesh sequence of the body dynamics and the corresponding mesh sequence of garment dynamics. Then,

To get the required information and saved in the folder of '/uv_body/' and '/uv_garment/', please refer the code in ./uv_rasterRender

To get the sign distance field of the body shape at conanical post, please refer to IGR: Implicit Geometric Regularization for Learning Shapes.

To get the sequence in '/garment_pnva/', please refer to garment_Data_Prep3.py

To get the sequence in '/body_RTVN/', please refer to body_Data_Prepare.py

To train the static generative garment deformation network, please refer to train_static_reconstruction.py

To get the garment correponding latent code sequence, please refer to run_static_reconstruction.py

To train the dynamic encoder, please refer to train_dynamic_prediction.py

To rollout the dynamic prediction with collision handling, please refer to run_dynamic_prediction.py

We also provide some pretrained checkpoint. You can play our code with our provided data, here.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published