Skip to content

laoreja/face-identity-transformer

Repository files navigation

Password-conditioned Face Identity Transformer

teaser

This is the PyTorch implementation of our paper Password-conditioned Anonymization and Deanonymization with Face Identity Transformers in ECCV 2020.

Abstract. Cameras are prevalent in our daily lives, and enable many useful systems built upon computer vision technologies such as smart cameras and home robots for service applications. However, there is also an increasing societal concern as the captured images/videos may contain privacy-sensitive information (e.g., face identity). We propose a novel face identity transformer which enables automated photo-realistic password-based anonymization and deanonymization of human faces appearing in visual data. Our face identity transformer is trained to (1) remove face identity information after anonymization, (2) recover the original face when given the correct password, and (3) return a wrong— but photo-realistic—face given a wrong password. With our carefully designed password scheme and multi-task learning objective, we achieve both anonymization and deanonymization using the same single network. Extensive experiments show that our method enables multimodal password conditioned anonymizations and deanonymizations, without sacrificing privacy compared to existing anonymization methods.

An overview of our network architecture: Network Architecture

Demo

Here is a demo of our method applied to video in the wild: Video demo

Environment setup

Our code is developed and tested under

  • Linux.
  • Python 3.6.
  • PyTorch 1.1 and torchvision 0.3.0. It works with PyTorch versions up to 1.3. For later versions, please notice PyTorch's breaking changes. In general, it should work under later versions of PyTorch.
  • NVIDIA GPU and CUDA is required. (With a batch size of 12 per GPU, the training fits ~12GB GPUs including GeForce GTX 1080 Ti.)

Get started

git clone https://github.com/laoreja/face-identity-transformer
cd face-identity-transformer
pip install -r requirements.txt
mkdir pretrained_models

Please download and extract the pretrained SphereFace model from the SphereFace PyTorch repo and put it under the directory pretrained_models.

Datasets

In our paper, we use the following three face datasets:

  • CASIA-Webface dataset.
  • LFW dataset: Download and extract from this link.
  • FFHQ dataset: We use the validation set of thumbnail version, since it matches the image size we use (128x128).

We use MTCNN to obtain the face bounding boxes and landmarks and you can download our extracted face bounding boxes and landmarks from this Google Drive folder. (The landmarks are only needed during training. Bounding boxes are needed for training and for images/video in the wild to crop out the face region). Please unzip the zipped txt file.

We store and organize all face datasets in the same directory data_root, and use an argument data_root in our code to fetch the datasets. Please download the datasets and set the data_root argument either in the command line or in the YAML config.

The data_root directory is organized as:

data_root
+-- CASIA-WebFace
+-- casia_landmark.txt
+-- lfw
+-- lfw_landmark.txt
+-- ffhq-dataset
|   +-- validation_paths.txt

Training

Adjust the batch_size in config/main.yaml and CUDA_VISIBLE_DEVICES according to the number of your available GPUs (12 x num_gpus).

We only use the CASIA-Webface for training. Please set the dataset argument to CASIA (current default setting).

CUDA_VISIBLE_DEVICES=0,1,2,3 python3 main_train.py config/main.yaml

Under the checkpoint directory args.ckpt_dir, the checkpoints, log, YAML config will be saved. You could also set up a http server (python3 -m http.server [PORT]) to check the intermediate results. web contains training visualizations and test_web contains validation visualizations.

During training, a visdom server will be automatically set up and you can check the training visualizations and loss curves there. (The default visdom port is 8097.)

Inference

Qualitative results on test images

CUDA_VISIBLE_DEVICES=0,1,2,3 python3 save_visual.py CHECKPOINT_DIR_TO_RESUME/YAML_FILE_NAME --ckpt_name checkpoint_xxxx.pth.tar --dataset DATASET [--inference_full] 

checkpoint_xxxx.pth.tar is the model checkpoint you would like to use for testing.

During training, our code automatically copies the config yaml file to the checkpoint directory (CHECKPOINT_DIR_TO_RESUME), and it would use the same config during inference time. The code would try to find the checkpoint with name checkpoint_xxxx.pth.tar from the directory CHECKPOINT_DIR_TO_RESUME.

You could add option --inference_full to do inference on the full test set, otherwise it would save visuliazations for one randomly sampled batch of test images.

The code would automatically set the batch size (batch_size_per_gpu_during_training x num_gpus_during_inference) according to the number of GPUs you use during inference from the CUDA_VISIBLE_DEVICES environment variable.

The qualitative results will be saved under the directory face-identity-transformer/qualitative_results.

Multimodality: Same input face with different passwords

Given different passwords, our approach achieves multimodality on the same input face, like the following illustration.

multimodality

CUDA_VISIBLE_DEVICES=0,1,2,3 python3 main_test_diff_passwds.py CONFIG_PATH --resume CHECKPOINT_PATH

CHECKPOINT_PATH is the path to the checkpoint (.pth.tar) file. Please make sure that the YAML config file in the arguments is the same config you used during training.

The results will be saved under the same parent directory as the CHECKPOINT_PATH. Assume CHECKPOINT_FOLDER_NAME is the folder name for the resumed checkpoint, then a new folder named test_CHECKPOINT_FOLDER_NAME will be created and results will be saved there, under the test_web subfoler. You could view the results through a http server. Each column in the index_latest.html represnets a different password (please ignore the column title there. Since we use a 16-bit password, sometimes the identity difference is subtle.)

Trained model

You can download our trained model from this Google drive folder. It also comes with the YAML config so that you could try save_visual.py with this folder.

We noticed some blob-like artifacts in generated images when the background is white. We found that decresing the learning rate for the discriminator and reduce the magnitude for each element of the passwords would help. We provide its YAML config file in config/one_fourth_D_lr_half.yaml and a pretrained model in this Google drive folder

Example Results

On CASIA dataset

: original image, : anonymized faces using different passwords, : recovered faces with correct passwords, : recovered faces with wrong passwords. Results on CASIA

Citation

If you find our work useful in your research, please cite our paper.

@inproceedings{gu2019password,
  title={Password-conditioned Anonymization and Deanonymization with Face Identity Transformers},
  author={Gu, Xiuye and Luo, Weixin and Ryoo, Michael S and Lee, Yong Jae},
  booktitle={European Conference on Computer Vision},
  year={2020}
}

Acknowledgements

We adapt clcarwin's SphereFace PyTorch implementation and use their pretrained model. Some of our code are adapted from CycleGAN and pix2pixHD implementations.

We also thank Jason Zhongzheng Ren for his support on this work, and our work is inspired from this awesome paper: Learning to Anonymize Faces for Privacy Preserving Action Detection, Zhongzheng (Jason) Ren, Yong Jae Lee, Michael S. Ryoo, in ECCV 2018.

About

The PyTorch implementation of our paper "Password-conditioned Anonymization and Deanonymization with Face Identity Transformers" in ECCV 2020.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages