End of training
Browse files- README.md +21 -1
- adapter_model.bin +1 -1
README.md
CHANGED
@@ -2,6 +2,7 @@
|
|
2 |
license: apache-2.0
|
3 |
library_name: peft
|
4 |
tags:
|
|
|
5 |
- generated_from_trainer
|
6 |
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
|
7 |
model-index:
|
@@ -111,7 +112,9 @@ fsdp_config:
|
|
111 |
|
112 |
# mixtral-lora
|
113 |
|
114 |
-
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on
|
|
|
|
|
115 |
|
116 |
## Model description
|
117 |
|
@@ -144,6 +147,23 @@ The following hyperparameters were used during training:
|
|
144 |
- lr_scheduler_warmup_steps: 10
|
145 |
- num_epochs: 2
|
146 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
147 |
### Framework versions
|
148 |
|
149 |
- PEFT 0.8.2
|
|
|
2 |
license: apache-2.0
|
3 |
library_name: peft
|
4 |
tags:
|
5 |
+
- axolotl
|
6 |
- generated_from_trainer
|
7 |
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
|
8 |
model-index:
|
|
|
112 |
|
113 |
# mixtral-lora
|
114 |
|
115 |
+
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on the None dataset.
|
116 |
+
It achieves the following results on the evaluation set:
|
117 |
+
- Loss: 0.1746
|
118 |
|
119 |
## Model description
|
120 |
|
|
|
147 |
- lr_scheduler_warmup_steps: 10
|
148 |
- num_epochs: 2
|
149 |
|
150 |
+
### Training results
|
151 |
+
|
152 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
153 |
+
|:-------------:|:-----:|:----:|:---------------:|
|
154 |
+
| 3.397 | 0.0 | 1 | 3.2822 |
|
155 |
+
| 0.1294 | 0.2 | 67 | 0.2029 |
|
156 |
+
| 0.1664 | 0.4 | 134 | 0.1918 |
|
157 |
+
| 0.1742 | 0.6 | 201 | 0.1853 |
|
158 |
+
| 0.163 | 0.8 | 268 | 0.1827 |
|
159 |
+
| 0.1537 | 1.0 | 335 | 0.1798 |
|
160 |
+
| 0.1056 | 1.19 | 402 | 0.1781 |
|
161 |
+
| 0.1688 | 1.39 | 469 | 0.1765 |
|
162 |
+
| 0.1187 | 1.59 | 536 | 0.1752 |
|
163 |
+
| 0.1823 | 1.79 | 603 | 0.1748 |
|
164 |
+
| 0.1022 | 1.99 | 670 | 0.1746 |
|
165 |
+
|
166 |
+
|
167 |
### Framework versions
|
168 |
|
169 |
- PEFT 0.8.2
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 969596450
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e26eed223bbbbaefc98dd7881d82cf2636f97ef84b71fee3376696995f7eb6ce
|
3 |
size 969596450
|