Skip to content

Implementation of Decoder only Transformers for experimentation

Notifications You must be signed in to change notification settings

mauicv/transformers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

38 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Transformers

Description:

This repo is a collection of PyTorch implementations of Transformer architectures with simple flexible config for ease of experimentation. The goal is learning and experimentation.

Tests:

Tests can be run using pytest from the root directory. There are also online colabs that should test any new architecture added to the repo on shakespeare character prediction.

  1. basic transformer
  2. MoE transformer
  3. Relative Attention

As well as this each architecture and layer should be benchmarked for speed using:

  1. Transformer-benchmarks
  2. Runtime Comparison

Resources:

  1. Attention Is All You Need
  2. On Layer Normalization in the Transformer Architecture
  3. minGPT
  4. The Annotated Transformer
  5. d2l-vision-transformer
  6. vector-quantize-pytorch

About

Implementation of Decoder only Transformers for experimentation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published