|
--- |
|
tags: |
|
- GEMMA |
|
- instruct |
|
- finetune |
|
- chatml |
|
- gpt4 |
|
- synthetic data |
|
- distillation |
|
model-index: |
|
- name: google/gemma-2-2b |
|
results: [] |
|
license: apache-2.0 |
|
language: |
|
- en |
|
library_name: transformers |
|
datasets: |
|
- teknium/OpenHermes-2.5 |
|
--- |
|
|
|
# Model Card for GEMMA2-2B-openhermes-2.5 |
|
|
|
This model is a fine-tuned version of Gemma 2 -2B on the OpenHermes-2.5 dataset. |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
This is a fine-tuned version of the google/gemma-2-2b model, trained on the OpenHermes-2.5 dataset. It is designed for instruction following and general language tasks. |
|
|
|
- **Developed by:** artificialguybr |
|
- **Model type:** Causal Language Model |
|
- **Language(s):** English |
|
- **License:** apache-2.0 |
|
- **Finetuned from model:** google/gemma-2-2b |
|
|
|
### Model Sources |
|
|
|
- **Repository:** https://huggingface.co/artificialguybr/Gemma2-2B-OpenHermes2.5 |
|
|
|
## Uses |
|
|
|
This model can be used for various natural language processing tasks, particularly those involving instruction following and general language understanding. |
|
|
|
### Direct Use |
|
|
|
The model can be used for tasks such as text generation, question answering, and other language-related applications. |
|
|
|
### Out-of-Scope Use |
|
|
|
The model should not be used for generating harmful or biased content. Users should be aware of potential biases in the training data. |
|
|
|
## Training Details |
|
|
|
### Training Data |
|
|
|
The model was fine-tuned on the teknium/OpenHermes-2.5 dataset. |
|
|
|
### Training Procedure |
|
|
|
#### Hardware and Software |
|
|
|
- **Hardware:** NVIDIA A100-SXM4-80GB (1 GPU) |
|
- **Software Framework:** 🤗 Transformers, Axolotl |
|
|
|
## Limitations and Biases |
|
|
|
More information is needed about specific limitations and biases of this model. |
|
|