Skip to content

Open-source Gloriously Extensive Yaml-configuration Repository for Reimplementing Architectures of Transformers

License

Notifications You must be signed in to change notification settings

TearGosling/OGEYRRAT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Open-source Gloriously Extensive Yaml-configuration Repository for Reimplementing Architectures of Transformers

(ogey rrat)

What the hell is this?

Here is a personal repo open to the public that is dedicated to me learning all types of Transformers by looking at other people's code and research papers and pretending that I know what I'm doing when I copy them and combine them together into an unholy abomination of build-your-own-Transformer. Basically, lucidrains' x-transformers library, but taken quite a few steps further by implementing training loops and supporting different libraries such as HuggingFace and ColossalAI

Why?

People say the best way of learning AI is to play with it yourself. That's what I'm doing here. Eventually, this might become a cool thing where all you have to do is specify a .yml file (either preset or custom-made) and in one command, begin to train it.

This just sounds like a bunch of autism.

That's exactly what this is.

To-do list:

  • Implement boilerplate code
  • Implement YAML-based config system (use OmegaConf library?)
  • Implement basic transformer architecture
  • Implement what is needed to create a basic training loop
    • Feed in datasets
      • HuggingFace support
      • PyTorch DataLoader support
    • Implement tokenizer(s)
    • Implement optimizer(s)
    • Actual process for training (zero_grad, forward, loss, backward, step)
    • Save checkpoints
  • Implement inferring from a model
  • Implement fine-tuning script

And more which I have probably forgotten.

About

Open-source Gloriously Extensive Yaml-configuration Repository for Reimplementing Architectures of Transformers

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages