Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is the 13B - unmodified Megatron gpt2 - baseline available? ( tr1-13B-base) #21

Open
ViktorThink opened this issue Dec 4, 2021 · 1 comment

Comments

@ViktorThink
Copy link

I was super excited to hear about this project! I was wondering if the model is available anywhere?

In the chronicles of tr1-13B-base it says at the end: "All checkpoints converted to HF format and uploaded to HUB.", which I thought meant that it is available on Huggingface, but I can't seem to find it.

Is it available and I'm just not able to find it, or did I misunderstand and it's not available?

@malteos
Copy link

malteos commented Apr 22, 2022

The checkpoint for the 1B version is available on HF Hub: https://huggingface.co/bigscience/tr5b-1B3-multilingual-alpha-checkpoints/tree/global_step118500 (Note: You must the select the branch depending on the global step - main branch is empty).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants