Skip to content

shatz01/slow-transformers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

20 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

๐ŸŒ slow-transformers

Our motto: "Go transformers! But dont go too fast. You still have to enjoy life โ˜ฎ๏ธ"

Diffability, noun

A principle underscoring the art of unmasking subtle divergences amidst complex similarities, diffability illuminates clear paths through intellectual labyrinths, providing clarity in a sea of cerebral complexity ... In practical terms: Understand the difference between two methods by diffing their code them in vscode.

Install

git clone ...
cd slow-transformers/
pip install -r requirements.txt

Supported Models

  • ViT
  • SimpleViT
  • Language Classification Transformer
  • Encoder-decoder model (generative)

Supported Datasets

  • cifar
  • imdb

TODO / Goals list

  • Vanilla transformer (or some language tasks)
  • fsdp/deepspeed
  • cross attention
  • more interesting architechtures (t5, perciever)
  • flash attention integration
  • jax?
  • resnet & hyena for comparison???
  • support m1
  • a script to run every model on every possible dataset and record everything in wandb (use hf trainer though)
  • also put datasets/dataloading entirely in file (move cifar from ./data to slow_vit.py, similar to hw_vit.py)

About

๐ŸŒ slow transformers

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages