Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Experiments] Various experiments related to MindGPT #67

Open
wants to merge 27 commits into
base: develop
Choose a base branch
from

Conversation

dudeperf3ct
Copy link
Member

@dudeperf3ct dudeperf3ct commented Aug 22, 2023

In this PR, we add 4 notebooks each focusing on a different experiment. Each notebook is a fully reproducible experiment environment.

  • big_models_experiment.ipynb: This notebook compares responses for NHS and Mind across various LLM models. Starting with flan-t5-base and keeping the rest of the variables constant, 5 additional LLM models were tested flan-t5-large, flan-t5-xl, fastchat-t5, openllam-3b and redpajama-3b-instruct.
  • Data Chunking: There are two parts of this experiments captured in two separate notebook
    • data_chunking_experiments.ipynb: This notebook handpick small subset of dataset relevant for the question and experiments with different sizes of chunk_size and chunk_overlap.
    • data_chunking_experiments_third_version_openllm.ipynb: This notebook takes similar approach above where instead of handpicking datapoints entire dataset (the latest version) is used. Two additional improvements have been added i.e. flan-t5-xl LLM and advanced template is used.
  • prompt_engineering_experiment.ipynb: This notebook experiments with 5 different prompts for 3 LLMs ( flan-t5-base , flan-t5-large and flan-t5-xl).

@dudeperf3ct dudeperf3ct self-assigned this Aug 22, 2023
@dudeperf3ct dudeperf3ct added the enhancement New feature or request label Aug 22, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

1 participant