Skip to content

POSTECH-CVLab/DeepVirtualMarkers

 
 

Repository files navigation

Deep Virtual Markers

This repository contains the accompanying code for Deep Virtual Markers for Articulated 3D Shapes, ICCV'21

| Paper | Video |

Athor Information

News

  • 2021-08-16: The first virsion of Deep Virtual Markers is published

Overview

We propose deep virtual markers, a framework for estimating dense and accurate positional information for various types of 3D data. We design a concept and construct a framework that maps 3D points of 3D articulated models, like humans, into virtual marker labels. To realize the framework, we adopt a sparse convolutional neural network and classify 3D points of an articulated model into virtual marker labels. We propose to use soft labels for the classifier to learn rich and dense interclass relationships based on geodesic distance. To measure the localization accuracy of the virtual markers, we test FAUST challenge, and our result outperforms the state-of-the-art. We also observe outstanding performance on the generalizability test, unseen data evaluation, and different 3D data types (meshes and depth maps). We show additional applications using the estimated virtual markers, such as non-rigid registration, texture transfer, and realtime dense marker prediction from depth maps.

Getting Started

Prerequisites

  • Ubuntu 18.06 or higher
  • CUDA 10.2 or higher
  • pytorch 1.6 or higher
  • python 3.8 or higher
  • GCC 6 or higher

Environment Setup

  • We recommend using docker
docker pull min00001/cuglmink

Demo Code (Docker)

Get sample data and pre-trained weight from here

docker pull min00001/cuglmink
./run_dvm_test.sh

License

This software is being made available under the terms in the LICENSE file.

Any exemptions to these terms requires a license from the Pohang University of Science and Technology.

Useful Links

Citing Deep Virtual Markers

@inproceedings{kim2021deep,
  title={Deep Virtual Markers for Articulated 3D Shapes},
  author={Hyomin Kim, Jungeon Kim, Jaewon Kam, Jaesik Park and Seungyong Lee},
  booktitle={ICCV},
  year={2021}
}

Related projects

NOTE : Our implementation is based on the "4D-SpatioTemporal ConvNets" repository

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 99.8%
  • Shell 0.2%