Skip to content

drishyakarki/nanoGPT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

nanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs. This repository contains code for a Transformer-based language model, specifically the Generative Pre-trained Transformer (GPT) model. GPT is a state-of-the-art language model architecture that has achieved impressive results in various natural language processing tasks, including text generation and language understanding. It is directly inspired by Andrej Karpathy's gpt video.

This repository contains the code that summed up from the Karpathy's makemore series, ultimately leading to this. I have implemented the code from makemore series too. You can check it out here: Makemore series

Requirements

simply, install the dependencies using

pip install -r requirements.txt

Training

To train the GPT model, follow these steps:

  1. Install the required dependencies: pip install -r requirements.txt
  2. Prepare your training data in a text file (data/train.txt).
  3. Run the training script:
cd nanoGPT
python gpt.py

You will also get sample outputs once the training script is completed.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published