Fill-Mask
Transformers
PyTorch
Guaraní
bert
gn-bert-base-cased / README.md
mmaguero's picture
Update README.md
3e9fdaf
|
raw
history blame
768 Bytes
---
language: gn
license: mit
datasets:
- wikipedia
- wiktionary
widget:
- text: "Paraguay ha'e peteĩ táva oĩva [MASK] retãme"
- text: "Augusto Roa Bastos ha'e peteĩ [MASK] arandu"
metrics:
- f1
- accuracy
---
# BERT-i-base-cased (gnBERT-base-cased)
A pre-trained BERT model for **Guarani** (12 layers, cased). Trained on Wikipedia + Wiktionary (~800K tokens).
# How cite?
```
@article{aguero-et-al2023multi-affect-low-langs-grn,
title={Multidimensional Affective Analysis for Low-resource Languages: A Use Case with Guarani-Spanish Code-switching Language},
author={Agüero-Torales, Marvin Matías, López-Herrera, Antonio Gabriel, and Vilares, David},
journal={Cognitive Computation},
year={2023},
publisher={Springer},
notes={Forthcoming}
}
```