Skip to content

How to fine-tune transformer models for text classification using Hugging Face Transformers and Datasets

License

Notifications You must be signed in to change notification settings

SMMousaviSP/huggingface_transformers_tutorial

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Fine-Tuning Hugging Face Transformers for Text Classification

This tutorial is based on Hugging Face course.

You can see the YouTube video recorded for this tutorial (Persian)

distilbert-base-uncased is recommended, since it's faster than bert-base-uncased and offers a good performance. Also it was pretrained with the same corpus as BERT. This model is aimed at being fine-tuned for NLP tasks such as text classification, token classification, and question answering, for text generation you should go for models such as gpt2. More information about this model is available here.

sst2 from glue benchmark is used on this tutorial. It consists of sentences from movie reviews and human annotations of their sentiment. The task is to predict the sentiment of a given sentence. It uses the two-way (positive/negative) class split, with only sentence-level labels.

Releases

No releases published

Packages

No packages published