Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Generation details for models under data/labeled #31

Open
caiqizh opened this issue Nov 19, 2023 · 1 comment
Open

Generation details for models under data/labeled #31

caiqizh opened this issue Nov 19, 2023 · 1 comment

Comments

@caiqizh
Copy link

caiqizh commented Nov 19, 2023

Thank you for the excellent work!

I have a question regarding the generation of outputs in the data/labeled files. Specifically, I'm curious about the parameters and prompts you used during this process. I've noticed that my generated text (e.g. from ChatGPT) is much longer than the content in your file. Could you please provide information on the settings you employed, such as temperature, max_tokens, and prompts, when generating the biographies? Your assistance in this matter would be greatly appreciated.

Thank you in advance!

@shmsw25
Copy link
Owner

shmsw25 commented Nov 20, 2023

Hi @caiqizh, thank you for your interest in our work.

Here is the prompt we used for ChatGPT:
image

Here are two hyperparameters:

  • temp=0.7 for both ChatGPT and InstructGPT
  • max_tokens=512 for InstructGPT and max_tokens=1024 for ChatGPT

Using different max_tokens should not affect the generations unless the generation exceed max_tokens , which never happened in our case. Given this, I think it is possible that you are seeing much longer responses due to the internal changes in ChatGPT (if it's not due to the difference in the prompt).

Let me know if you have any further questions. Thanks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants