Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,8 @@ datasets:
|
|
10 |
language:
|
11 |
- en
|
12 |
base_model: trollek/LittleInstructionMaker-4B-v0.2
|
|
|
|
|
13 |
---
|
14 |
# Tyr-4B-v0.1
|
15 |
|
@@ -19,10 +21,15 @@ Merging seems like the way to go when it comes to training language models on a
|
|
19 |
|
20 |
## Model Description
|
21 |
|
22 |
-
This model was created by training LoRAs and
|
23 |
|
24 |
Incedentally, it seems uncensored. It was trained using the ChatML template and can be used with or without a system prompt.
|
25 |
|
|
|
|
|
|
|
|
|
|
|
26 |
## Apache-2.0 + WhiteRabbitNeo Extended Version
|
27 |
|
28 |
### Licence: Usage Restrictions
|
|
|
10 |
language:
|
11 |
- en
|
12 |
base_model: trollek/LittleInstructionMaker-4B-v0.2
|
13 |
+
tags:
|
14 |
+
- not-for-all-audiences
|
15 |
---
|
16 |
# Tyr-4B-v0.1
|
17 |
|
|
|
21 |
|
22 |
## Model Description
|
23 |
|
24 |
+
This model was created by training LoRAs ([rsLoRA](https://arxiv.org/abs/2312.03732) with [NEFTune](https://arxiv.org/abs/2310.05914) noise alpha at 5) and [DELLA merge](https://arxiv.org/abs/2406.11617) them. It saves space and time this way and the result is good. The WhiteRabbitNeo datasets are the focus in this one along with coding.
|
25 |
|
26 |
Incedentally, it seems uncensored. It was trained using the ChatML template and can be used with or without a system prompt.
|
27 |
|
28 |
+
I am very grateful for all the different components that made this model possible. From H2O and their danube models, through Huggingface and Llama-Factory for making the fine-tuning easy, and to all the great dataset creators. **Thank you!**
|
29 |
+
|
30 |
+
### Quants
|
31 |
+
- [trollek/Tyr-4B-v0.1-GGUF](https://huggingface.co/trollek/Tyr-4B-v0.1-GGUF)
|
32 |
+
|
33 |
## Apache-2.0 + WhiteRabbitNeo Extended Version
|
34 |
|
35 |
### Licence: Usage Restrictions
|