Skip to content

yanx27/JS3C-Net

Repository files navigation

JS3C-Net

Sparse Single Sweep LiDAR Point Cloud Segmentation via Learning Contextual Shape Priors from Scene Completion (AAAI2021)

This repository is for JS3C-Net introduced in the following AAAI-2021 paper [arxiv paper]

Xu Yan, Jiantao Gao, Jie Li, Ruimao Zhang, Zhen Li*, Rui Huang and Shuguang Cui, "Sparse Single Sweep LiDAR Point Cloud Segmentation via Learning Contextual Shape Priors from Scene Completion".

  • Semantic Segmentation and Semantic Scene Completion:

If you find our work useful in your research, please consider citing:

@inproceedings{yan2021sparse,
  title={Sparse Single Sweep LiDAR Point Cloud Segmentation via Learning Contextual Shape Priors from Scene Completion},
  author={Yan, Xu and Gao, Jiantao and Li, Jie and Zhang, Ruimao and Li, Zhen and Huang, Rui and Cui, Shuguang},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={35},
  number={4},
  pages={3101--3109},
  year={2021}
}

Getting Started

Set up

Clone the repository:

git clone https://github.com/yanx27/JS3C-Net.git

Installation instructions for Ubuntu 16.04:

  • Make sure CUDA and cuDNN are installed. Only this configurations has been tested:
    • Python 3.6.9, Pytorch 1.3.1, CUDA 10.1;
  • Compile the customized operators by sh complile.sh in /lib.
  • Install spconv1.0 in /lib/spconv. We use the same version with PointGroup, you can install it according to the instruction. Higher version spconv may cause issues.

Data Preparation

  • SemanticKITTI and SemanticPOSS datasets can be found in semantickitti-page and semanticposs-page.
  • Download the files related to semantic segmentation and extract everything into the same folder.
  • Use voxelizer generate ground truths of semantic scene completion, where following parameters are used. We provide pre-processed SemanticPOSS SSC labels here.
min range: 2.5
max range: 70
future scans: 70
min extent: [0, -25.6, -2]
max extent: [51.2, 25.6,  4.4]
voxel size: 0.2
  • Finally, the dataset folder should be organized as follows.
SemanticKITTI(POSS)
├── dataset
│   ├── sequences
│   │  ├── 00
│   │  │  ├── labels
│   │  │  ├── velodyne
│   │  │  ├── voxels
│   │  │  ├── [OTHER FILES OR FOLDERS]
│   │  ├── 01
│   │  ├── ... ...

  • Note that the data for official SemanticKITTI SSC benchmark only contains 1/5 of the whole sequence and they provide all extracted SSC data for the training set here.
  • (New) In this repo, we use old version of SemanticKITTI, and there is a bug of generating SSC data contains a wrong shift on upwards direction (see issue). Therefore, we add an additional shifting to align their old version dataset here, and if you use the newest version of data, you can delete it. Also, you can check the alignment ratio by using --debug.

SemanticKITTI

Training

Run the following command to start the training. Output (logs) will be redirected to ./logs/JS3C-Net-kitti/. You can ignore this step if you want to use our pretrained model in ./logs/JS3C-Net-kitti/.

$ python train.py --gpu 0 --log_dir JS3C-Net-kitti --config opt/JS3C_default_kitti.yaml

Evaluation Semantic Segmentation

Run the following command to evaluate model on evaluation or test dataset

$ python test_kitti_segment.py --log_dir JS3C-Net-kitti --gpu 0 --dataset [val/test]

Evaluation Semantic Scene Completion

Run the following command to evaluate model on evaluation or test dataset

$ python test_kitti_ssc.py --log_dir JS3C-Net-kitti --gpu 0 --dataset [val/test]

SemanticPOSS

Results on SemanticPOSS can be easily obtained by

$ python train.py --gpu 0 --log_dir JS3C-Net-POSS --config opt/JS3C_default_POSS.yaml
$ python test_poss_segment.py --gpu 0 --log_dir JS3C-Net-POSS

Pretrained Model

We trained our model on a single Nvidia Tesla V100 GPU with batch size 6. If you want to train on the TITAN GPU, you can choose batch size as 2. Please modify dataset_dir in args.txt to your path.

Model #Param Segmentation Completion Checkpoint
JS3C-Net 2.69M 66.0 56.6 18.5MB

Results on SemanticKITTI Benchmark

Quantitative results on SemanticKITTI Benchmark at the submisison time.

Acknowledgement

This project is not possible without multiple great opensourced codebases.

License

This repository is released under MIT License (see LICENSE file for details).