Skip to content

xcyan/nips16_PTN

Repository files navigation

Perspective Transformer Nets (PTN)

This is the code for NIPS 2016 paper Perspective Transformer Nets: Learning Single-View 3D Object Reconstruction without 3D Supervision by Xinchen Yan, Jimei Yang, Ersin Yumer, Yijie Guo and Honglak Lee

Please follow the instructions to run the code.

Requirements

PTN requires or works with

  • Mac OS X or Linux
  • NVIDIA GPU

Installing Dependency

The following command installs the Perspective Transformer Layer:

./install_ptnbhwd.sh

Dataset Downloading

  • Please run the command to download the pre-processed dataset (including rendered 2D views and 3D volumes):
./prepare_data.sh

Pre-trained Models Downloading (single-class experiment)

PTN-Proj: ptn_proj.t7

PTN-Comb: ptn_comb.t7

CNN-Vol: cnn_vol.t7

  • The following command downloads the pre-trained models:
./download_models.sh

Testing using Pre-trained Models (single-class experiment)

  • The following command evaluates the pre-trained models:
./eval_models.sh

Training (single-class experiment)

  • If you want to pre-train the view-point indepedent image encoder on single-class, please run the following command. Note that the pre-training could take a few days on a single TITAN X GPU.
./demo_pretrain_singleclass.sh
  • If you want to train PTN-Proj (unsupervised) on single-class based on pre-trained encoder, please run the command.
./demo_train_ptn_proj_singleclass.sh
  • If you want to train PTN-Comb (3D supervision) on single-class based on pre-trained encoder, please run the command.
./demo_train_ptn_comb_singleclass.sh
  • If you want to train CNN-Vol (3D supervision) on single-class based on pre-trained encoder, please run the command.
./demo_train_cnn_vol_singleclass.sh

Using your own camera

  • In many cases, you want to implement your own camera matrix (e.g., intrinsic or extrinsic). Please feel free to modify this function.

  • Before start your own implementation, we recommand to go through some basic camera geometry in this computer vision textbook written by Richard Szeliski (see Eq 2.59 at Page 53).

  • Note that in our voxel ray-tracing implementation, we used the inverse camera matrix.

Third-party Implementation

Besides our torch implementation, we recommend to see also the following third-party re-implementation:

  • TensorFlow Implementation: This re-implementation was developed during Xinchen's Google internship; If you find a bug, please file a bug including @xcyan.

Citation

If you find this useful, please cite our work as follows:

@incollection{NIPS2016_6206,
title = {Perspective Transformer Nets: Learning Single-View 3D Object Reconstruction without 3D Supervision},
author = {Yan, Xinchen and Yang, Jimei and Yumer, Ersin and Guo, Yijie and Lee, Honglak},
booktitle = {Advances in Neural Information Processing Systems 29},
editor = {D. D. Lee and M. Sugiyama and U. V. Luxburg and I. Guyon and R. Garnett},
pages = {1696--1704},
year = {2016},
publisher = {Curran Associates, Inc.},
url = {http://papers.nips.cc/paper/6206-perspective-transformer-nets-learning-single-view-3d-object-reconstruction-without-3d-supervision.pdf}
}

About

Torch Implementation of NIPS'16 paper: Perspective Transformer Nets

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published