Skip to content

chandar-lab/CGOptimizer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Critical Gradient Optimizers

Critical Gradient Optimizers from the "Memory Augmented Optimizers for Deep Learning" project and paper, reformatted as package and stripped down to just the necessary components to integrate the optimizers into your code.

This code is compatible with the following versions:

python >= 3.6
pytorch >= 1.7.1

Installation

Install these optimizers using

pip install cgoptimizer

Alternatively, clone this repository anywhere on your system. Once cloned, cd to the directory and install with: pip install .

Colab

The shared colab notebook shows examples on using the Critical gradient optimizers on toy classification tasks constructed with scikit-learn package.

Importing and Running

You can import the optimizers as you would any PyTorch optimizer. There are no requirements to run them other than PyTorch and its dependencies.

When installed, import the optimizers to your training script as needed:

from cgoptimizer import SGD_C, RMSprop_C, Adam_C, AdamW_C

You can then replace any PyTorch optimizer in your script with their _C counterpart. Note that currently only Critical-Gradient variants of Adam, RMSprop, and SGD (with optional momentum but NOT Nesterov) are implemented.

Here is a sample replacement:

optimizer = Adam(model.parameters(), lr=0.001)

becomes

optimizer = Adam_C(model.parameters(), lr=0.001, **kwargs)

Similarly, for efficient GPU-based implementation:

from cgoptimizer.optim_eff import SGD_C_eff, RMSprop_C_eff, Adam_C_eff, AdamW_C_eff
optimizer = Adam_C_eff(model.parameters(), lr=0.001, **kwargs)

Optimizer Usage and Tuning

The Critical Gradient variants use all the same hyperparameters as their vanilla counterparts, so you may not need to perform any additional tuning.

The _C (and _C_eff) optimizers have two additional hyperparameters compared to the vanilla version: topC which indicates how many critical gradients to keep anddecay which indicates how much the norms of critical gradients are decayed each step. These are keyword arguments with default values which we observed to work well. For additional performance, these can be tuned.

The _C (and _C_eff) variants perform best using either the same best learning rate as its vanilla counterpart, or 1/10 that learning rate. It is recommended you run both learning rates to compare.

Hyperparameter topC determines how many critical gradients are stored and thus how much memory is used. Higher topC usually result in longer training times. Good topC values usually fall between 5 and 20. We recommended using values 5, 10, and 20.

Hyperparameter decay indicates the level of decay in the buffer. This modifies how frequently the buffer is refreshed. The decay parameter must fall between 0 and 1. We recommended using values 0.7 and 0.9.

Citation

@misc{mcrae2021memory,
  author    = {McRae, Paul-Aymeric and Parthasarathi, Prasanna and Assran, Mahmoud and Chandar, Sarath},
  title     = {Memory Augmented Optimizers for Deep Learning},
  year      = {2022},
  booktitle = {Proceedings of ICLR}
}

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages