Skip to content
This repository has been archived by the owner on Mar 19, 2024. It is now read-only.

What config relate to learning rate warm up, weight decay, and momentum in 1 node n GPUs (n > 1 && n < 8) config? #584

Open
tungts1101 opened this issue Jan 30, 2024 · 0 comments

Comments

@tungts1101
Copy link

❓ How to do something using VISSL

Describe what you want to do, including:

  1. what I am trying to do: I have read the paper Imagenet-1hour. In there they mentioned the learning rate warm-up, weight decay, and momentum when implementing distributed training in 1 node multi gpus. However, I could not find any documents related to these configs. How could I properly set them?
  2. what outputs you are expecting: A config and an explanation related to learning rate warm-up strategy, weight decay, and momentum in 1 node n gpus machine?

❓ What does an API do and how to use it?

Please link to which API or documentation you're asking about from
https://github.com/facebookresearch/vissl/tree/main/docs

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant