Skip to content

D-Lab's 2-hour introduction to Generative Pretrained Transformers (GPT) for beginners. Learn about text encoding, word embeddings, and the transformer architecture upon which GPT is based. Create texts using a GPT model with the Transformers library in Python, and learn about hyperparameters such as temperature.

dlab-berkeley/GPT-Fundamentals

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

55 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

D-Lab's GPT Fundamentals Workshop

DataHub Binder License: CC BY 4.0

This repository contains the materials for D-Lab's GPT Fundamentals workshop.

Prerequisites

No technical background is required, but we recommend attending Python Fundamentals prior to this workshop in order to understand running Python code cells in Jupyter notebooks.

Check out D-Lab's Learning Pathways to figure out which of our workshops to take!

Workshop Goals

This workshop offers a general introduction to the GPT (Generative Pretrained Transformers) model. We will explore the transformer architecture upon which GPT models are built, how transformer models encode natural language into embeddings, and how GPT predicts text.

Learning Objectives

After this workshop, you will be able to:

  • Understand what makes the "magic" of GPT and Large Language Models (LLMs) possible via foundational concepts of word embeddings and the transformer architecture.
  • Get an intuition of word embeddings as numbers projected in vector space.
  • Distinguish components of the self-attention mechanism in transformer models.
  • Adjust hyperparameters of GPT models to control output randomness.
  • Use geometric similarity approaches to represent varying types of similarity between words.

This workshop does not cover the following:

  • Prompt Engineering.
  • RLHF finetuning.
  • The fundamentals of Python. See the list of workshops below to get a better grasp on the Python language used in this workshop.

Installation Instructions

We will use Python to go through the workshop materials. Complete the following steps:

  1. Install the Anaconda distribution of Python on your machine.
  2. Download these workshop materials:
    • Click the green "Code" button in the top right of the repository information.
    • Click "Download Zip".
    • Extract this file to a folder on your computer where you can easily access it (we recommend Desktop).
  3. Optional: if you’re familiar with git, you can instead clone this repository by opening a terminal and entering git clone https://github.com/dlab-berkeley/GPT-Fundamentals.git

Minimum Specifications:

  • Processor: At least a modern quad-core processor (i5 or i7). More cores are beneficial for parallel processing.
  • RAM: 8 GB is the bare minimum, but 16 GB or more is recommended, especially for larger models.
  • Storage: SSD (Solid State Drive) is preferred for faster data reading and writing.
  • Operating System: Linux or Windows with Python environment set up.

Ideal Specifications:

  • Processor: High-end i7 or i9, or an equivalent AMD Ryzen processor.
  • RAM: 16 GB or more.
  • GPU: A dedicated GPU with CUDA support is highly beneficial. Models like NVIDIA RTX 2060 or better can significantly speed up computation. The larger the model, the more VRAM is needed. * For the largest GPT-2 model, a GPU with at least 8 GB of VRAM is recommended.
  • Storage: SSD with sufficient space for the model and your datasets.

Is Python not Working on Your Computer?

If you do not have Python installed and the materials loaded on your workshop by the time it starts, we strongly recommend using the UC Berkeley Datahub to run the materials for these lessons. You can access the DataHub by clicking the following button:

DataHub

The DataHub downloads this repository, along with any necessary packages, and allows you to run the materials in an RStudio instance on UC Berkeley's servers. No installation is necessary from your end - you only need an internet browser and a CalNet ID to log in. By using the DataHub, you can save your work and come back to it at any time. When you want to return to your saved work, just go straight to the D-Lab DataHub, sign in, and you click on the GPT-Fundamentals folder.

If you don't have a Berkeley CalNet ID, you can still run these lessons in the cloud, by clicking this button:

Binder

By using this button, however, you cannot save your work.

Other D-Lab Python Workshops

Here are other Python workshops offered by the D-Lab:

Introductory Workshops

Advanced Workshops

About the UC Berkeley D-Lab

D-Lab works with Berkeley faculty, research staff, and students to advance data-intensive social science and humanities research. Our goal at D-Lab is to provide practical training, staff support, resources, and space to enable you to use R for your own research applications. Our services cater to all skill levels and no programming, statistical, or computer science backgrounds are necessary. We offer these services in the form of workshops, one-to-one consulting, and working groups that cover a variety of research topics, digital tools, and programming languages.

Visit the D-Lab homepage to learn more about us. You can view our calendar for upcoming events, learn about how to utilize our consulting and data services, and check out upcoming workshops.

Contributors

Tom van Nuenen

Renata Barreto

About

D-Lab's 2-hour introduction to Generative Pretrained Transformers (GPT) for beginners. Learn about text encoding, word embeddings, and the transformer architecture upon which GPT is based. Create texts using a GPT model with the Transformers library in Python, and learn about hyperparameters such as temperature.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published