Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

eval opt-175B #63

Open
henan991201 opened this issue Oct 26, 2022 · 1 comment
Open

eval opt-175B #63

henan991201 opened this issue Oct 26, 2022 · 1 comment

Comments

@henan991201
Copy link

I noticed you evaluated the opt-175B model, how did it convert to a Megatron-Deepspeed checkpoint? I can not find a 175B huggingface transformers checkpoint. Also, I can not successfully convert the opt-66B checkpoint. @thomasw21 Thanks for any reply!

@thomasw21
Copy link
Member

Hi, we used ranking evaluation for opt-175b, ie getting the loglikelihood of examples. Typically a classification task would look like:

sample1: "Review: {REVIEW}\nIs this review positive or negative? positive
sample2: "Review: {REVIEW}\nIs this review positive or negative? negative

We compute the likelihood of both samples, and rank them to get accuracy/recall/f1 for example.

Concerning the way we converted Megatron-DeepSpeed checkpoint, we essentially computed the opposite of https://github.com/huggingface/transformers/blob/main/src/transformers/models/bloom/convert_bloom_original_checkpoint_to_pytorch.py typically mapped each parameters to their stages (in the DeepSpeed sense), you can find:

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants