Skip to content

Latest commit

 

History

History
31 lines (24 loc) · 1.54 KB

README.md

File metadata and controls

31 lines (24 loc) · 1.54 KB

gpt2-fastapi-hugingface

This is an example how to do fine tuning with your own datasets from huggingface gpt2 models

Installation

  1. git clone this repository

  2. pip install -r requirements.txt

  3. FYI Huggingface only allowing 1024byte of data on every training, so we should chunk our data
    run prepare_data.py to generate got.txt into chunked

  4. run models.py to fine tuned our model into existing huggingface data
    Turn on the api by typing this

  5. uvicorn main:app --reload

Hit api

To hit the api you can open postman/any other apps for api post you want

hit into localhost:8000/generate
with this payload as json

{"prefix": "The young knight",
"max_length": 800,
"top_k": 5 }

To train it faster

If your machine is equipped with CUDA capabilities, I kindly request that you uncomment a few lines in the models.py and runner.py files. These lines pertain to CUDA-specific configurations and optimizations. However, please note that I do not have access to a CUDA-enabled machine, so I have disabled these lines in my version of the code.

To ensure compatibility and seamless execution on non-CUDA machines, I have thoroughly tested the project without CUDA dependencies. Rest assured that the functionality and performance remain unaffected.

Code Explanation is on my medium

Click here for code explanation