doken-tokuyama / traingpt2-127m-fromscratch Goto Github PK
View Code? Open in Web Editor NEWThis project forked from norod/traingpt2-127m-fromscratch
A trio of Google-Colab notebooks (ipynb) for training a GPT-2 (127M) model from scratch (useful for other / non-English languages) using gpt-2-simple