Safetensors

CUTE-Llama

CUTE-Llama 是一个基于 Llama2-7B 的多语言模型,专门针对中文、维吾尔语和藏语进行了优化。该模型是 CUTE (Chinese, Uyghur, Tibetan, English) Dataset 项目的一部分,旨在增强低资源语言的跨语言知识迁移能力。

模型版本

我们提供两个版本的模型:

  • CUTE-Llama-Parallel: 使用平行语料训练的版本
  • CUTE-Llama-Non-Parallel: 使用非平行语料训练的版本

相关链接

引用

如果您使用了我们的模型,请引用我们的论文:

@inproceedings{zhuang2025cute,
  title={CUTE: A Multilingual Dataset for Enhancing Cross-Lingual Knowledge Transfer in Low-Resource Languages},
  author={Zhuang, Wenhao and Sun, Yuan},
  booktitle={Proceedings of the 31st International Conference on Computational Linguistics},
  pages={10037--10046},
  year={2025}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for CMLI-NLP/CUTE-Llama

Finetuned
(783)
this model

Dataset used to train CMLI-NLP/CUTE-Llama