Skip to content

Code for the paper "Addressing caveats of neural persistence with deep graph persistence".

Notifications You must be signed in to change notification settings

ExplainableML/Deep-Graph-Persistence

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Deep Graph Persistence

This is the code for our work Addressing caveats of neural persistence with deep graph persistence.
Our paper has been accepted to TMLR. You can find it at the following locations:

Also, we have prepared a video presentation summarizing the main findings: https://youtu.be/KfCpoPYK_CY

Overview

This repository contains code for replicating our empirical experiments regarding neural persistence (NP) in trained neural networks. Also, this repository contains code for replicating our experiments on using deep graph persistence for detecting image corruptions.

For calculating NP, we rely on the great code from the original Neural Persistence: A Complexity Measure for Deep Neural Networks Using Algebraic Topology paper.

Setup

Besides standard dependencies available via pip, this repository also requires the Aleph library. Detailed instructions on how to install Aleph can be found in the corresponding repository.

After installing Aleph, the remaining dependencies can be installed from requirements.txt via

pip install -r requirement.txt

Replication

Analysis of Neural Persistence

In order to follow the instructions, first change to src/neural-persistence-analysis.

To train MLP models, run

python main.py

This will automatically save a model checkpoint every quarter epoch and also log relevant metrics in ./log. To specify experiment details, set the following parameters:

  • --hidden: The hidden size
  • --layers: The number of hidden layers
  • --activation: The activation function. Options are relu, tanh, or sigmoid
  • --dataset: The dataset to train the MLP on. Options are mnist, fashion-mnist, and emnist

The full list of parameters, run python main.py --help.

Weight-shuffling experiment

To replicate the weight-shuffling experiments, run

python get_permutation_analysis.py

and set the following parameters:

  • --path: Path to the model checkpoint
  • --output_path: Path where to store results
  • --num_permutations: How many times to shuffle weights and re-calculate neural persistence

Combining logs

We provide a script parse_logs.py that collects logs from different runs of main.py, get_permutation_analysis.py, and get_weight_distribution.py and saves them in single .csv files.

To use this script, adjust the file paths (path_to_logs, path_to_permutation_analysis, path_to_weight_distribution_analysis) in the file and run

python parse_logs.py

Detecting corrupted images with Deep Graph Persistence

To follow the instructions, first change to src/dgp-shift-detection. We also provide example scripts in src/dgp-shift-detection/example-scripts to illustrate running experiments in parallel using the slurm resource manager.

1. Train Models

Replicating this experiment involves several steps. The first step is to train MLP models required to calculate metrics such as DGP or Topological Uncertainty. To train MLP models, run

python train.py

while setting the desired parameters. The most important parameters are:

  • --hidden: The hidden size
  • --layers: The number of hidden layers
  • --dataset: The dataset to train the MLP on. Options are mnist, fashion-mnist, cifar10 as defined in load_data.py
  • --runs: How many replications with the same hyperparameters, but different initialization, we want to train
  • --checkpoint_root_dir: Where to save model checkpoints

A full list can be found by running python train.py --help.

2. Prepare data

The next step is preparing the data, i.e. the corrupted images. To prepare the data, run

python make_shifted_data.py

while setting the desired parameters. The relevant parameters are:

  • --dataset: The dataset to take images from. Options are mnist, fashion-mnist, cifar10 as defined in load_data.py
  • --num-train-samples: The number of images to use for training purposes. Recall that some methods, like Topological Uncertainty, require a number of image representations as reference
  • --num-test-samples: The number of samples to corrupt for testing purposes

3. Extract image representations

Then, we have to extract the representations of corrupted and non-corrupted (clean) images in various ways. To extract representations, run

python make_witness_vectors.py

while setting the appropriate parameters. The following parameters are required:

  • --dataset: The path to a dataset produced by make_shifted_data.py (has file ending .npy)
  • --model-name: Specifies the model's hyperparameters and must be of form data=$dataset-hidden=$hidden-layers=$layers-run=$run
  • --model-path: Path to a model checkpoint resulting from running train.py. Image representations are extracted from this model
  • --shift: The corruption to extract representations for. Options are gaussian_noise, salt_pepper_noise, gaussian_blur, image_transform, uniform_noise, pixel_shuffle, pixel_dropout, as defined in data_shifts.py
  • --intensity: The corruption intensity. Must be an integer between (including) 0 and 5
  • --method: The method used for extracting image representations. Options are: softmax, logit, magdiff, dgp, dgp_normalize, dgp_var, dgp_var_normalize, dgp_approximate, dgp_var_approximate, topological, topological_var, input_activation as defined in witness_functions.py

4. Calculate evaluation metrics

Finally, we evaluate the accuracy of detecting corrupted image (batches) by running

python magdiff_evaluation.py

where the following parameters are required:

  • --dataset
  • --hidden
  • --layers
  • --run
  • --shift
  • --intensity
  • --method

These parameters specify the MLP model and the type of corruption for which representations have been extracted by running make_witness_vectors.py. Results will be saved in a subdirectory of ./evaluation_results.

References

If you use this code, please cite

@article{
    girrbach2023addressing,
    title={Addressing caveats of neural persistence with deep graph persistence},
    author={Leander Girrbach and Anders Christensen and Ole Winther and Zeynep Akata and A. Sophia Koepke},
    journal={Transactions on Machine Learning Research},
    issn={2835-8856},
    year={2023}
}

About

Code for the paper "Addressing caveats of neural persistence with deep graph persistence".

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published