benchang1110's picture
Update README.md
543ede3 verified
|
raw
history blame
448 Bytes
metadata
library_name: transformers
license: apache-2.0
datasets:
  - benchang1110/pretrainedtw
  - HuggingFaceTB/cosmopedia-100k
language:
  - zh
widget:
  - text: 在很久以前,這座島上
    example_title: Example1

Model Card for Model ID

This is a continue-pretrained version of Tinyllama tailored for traditional Chinese. The continue-pretraining dataset contains roughly 2B tokens.