Skip to content

pretraining GPT 1 from scratch. Implementing Multihead Attention (MHA) using PyTorch from: Improving Language Understanding by Generative Pre-Training (https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf)

License

Notifications You must be signed in to change notification settings

MyDarapy/gpt-1-from-scratch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 

Repository files navigation

About

pretraining GPT 1 from scratch. Implementing Multihead Attention (MHA) using PyTorch from: Improving Language Understanding by Generative Pre-Training (https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf)

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages