Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,42 @@
|
|
1 |
---
|
|
|
2 |
library_name: peft
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
## Training procedure
|
5 |
|
6 |
|
|
|
1 |
---
|
2 |
+
license: agpl-3.0
|
3 |
library_name: peft
|
4 |
+
tags:
|
5 |
+
- not-for-all-audiences
|
6 |
---
|
7 |
+
|
8 |
+
# Pippa-13b-qlora
|
9 |
+
|
10 |
+
This is a repository of my Llama-2-13b Qlora checkpoints of the [PIPPA-13b-ShareGPT](https://huggingface.co/datasets/kingbri/PIPPA-ShareGPT) dataset.
|
11 |
+
|
12 |
+
You can read more about the dataset on its relevant page. It's a ShareGPT reformat of the [PIPPA dataset](https://huggingface.co/datasets/PygmalionAI/PIPPA) by PygmalionAI. The reformat was done to allow for axolotl compatability.
|
13 |
+
|
14 |
+
### Architecture
|
15 |
+
|
16 |
+
- **Model Architecture**: Llama-2-13b
|
17 |
+
- **Training Algorithm**: QLora
|
18 |
+
- **Dataset Used**: PIPPA-ShareGPT (pippa_sharegpt_trimmed.jsonl)
|
19 |
+
|
20 |
+
### Training Details
|
21 |
+
|
22 |
+
- **Dataset**: [PIPPA-ShareGPT](https://huggingface.co/datasets/kingbri/PIPPA-ShareGPT)
|
23 |
+
- **Datset type**: ShareGPT
|
24 |
+
- **Training Parameters**: [See Here](https://gist.github.com/bdashore3/55ae04892f31609f2c3779c4a8a55408)
|
25 |
+
- **Training Environment**: Axolotl
|
26 |
+
- **sequence_len**: 4096
|
27 |
+
|
28 |
+
## Notes
|
29 |
+
|
30 |
+
This Qlora was produced as an experiment to see how the public version of PIPPA can affect a model. As a result, I have no idea if this lora is of great quality or absolute garbage.
|
31 |
+
|
32 |
+
## Acknowledgments
|
33 |
+
|
34 |
+
Thanks to:
|
35 |
+
- PygmalionAI: The creators of the PIPPA dataset
|
36 |
+
- Axolotl: Finetuning suite
|
37 |
+
|
38 |
+
## Axolotl stuff
|
39 |
+
|
40 |
## Training procedure
|
41 |
|
42 |
|