Skip to content

Implementing a GPT (Generative Pre-trained Transformer) model from scratch on Shakespeare's work.

Notifications You must be signed in to change notification settings

JivanAcharya/Shakespeare-GPT

Repository files navigation

Shakespeare-GPT

Implementing a GPT (Generative Pre-trained Transformer) model from scratch, guided by the insightful tutorial from the visionary genius himself: Andrej Karpathy trained on Shakespeare's work.

About Model

  • About 10M parameters
  • Trained on Shakespeare's work
  • Implemented Self attention and Multiheaded Attention
  • Decoder only model

Running on Your Device

  • Clone the Github repo
  • Further Training can be done
  • For inference only need (model.py, model_state_dict.pt and streamlit.py files)
  • Run the streamlit interface with streamlit run streamlit.py (Inteface will look as shown below) Streamlit Implementation

About

Implementing a GPT (Generative Pre-trained Transformer) model from scratch on Shakespeare's work.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published