File size: 1,090 Bytes
44b40ec 738c2e5 44b40ec 8608e76 44b40ec 738c2e5 44b40ec 738c2e5 44b40ec 738c2e5 7546127 457bdb7 ec3909a 44b40ec ec3909a 44b40ec |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 |
---
tags:
- generated_from_trainer
model-index:
- name: gpt-est-base
results: []
widget:
- text: ">wiki< mis on GPT? Vastus:"
---
---
# gpt-est-base
This is the base-size [GPT2](https://huggingface.co/docs/transformers/model_doc/gpt2) model, trained from scratch on 2.2 billion words (Estonian National Corpus + News Crawl + Common Crawl) for 3 epochs. Previously named "gpt-4-est-base", renamed to avoid click-baiting.
[Reference](https://doi.org/10.22364/bjmc.2022.10.3.19)
### Format
For training data was prepended with a text domain tag, and it should be added as prefix when using the model: >general<, >web<, >news<, >doaj< and >wiki< (standing for general texts, web crawled texts, news, article abstracts and wikipedia texts). Use the prefixes like this, e.g: ">web< Kas tead, et".
### Model details
- num. of layers: 12
- num. of heads: 12
- embedding size: 768
- context size: 1024
- total size: 118.68M params
Further details to be added soon.
### Framework versions
- Transformers 4.13.0.dev0
- Pytorch 1.10.0+cu102
- Datasets 1.15.1
- Tokenizers 0.10.3
|