Commit
·
2de5db5
1
Parent(s):
8903c39
Update README.md
Browse filesThis is an extremely small 51M parameters only GPT-2 scaled down model trained on a small fraction of The Pile dataset (roughly 10 Billion Tokens).