PEFT
Not-For-All-Audiences
kingbri commited on
Commit
7aa2af3
·
1 Parent(s): 7f8e339

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -0
README.md CHANGED
@@ -1,6 +1,42 @@
1
  ---
 
2
  library_name: peft
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  ## Training procedure
5
 
6
 
 
1
  ---
2
+ license: agpl-3.0
3
  library_name: peft
4
+ tags:
5
+ - not-for-all-audiences
6
  ---
7
+
8
+ # Pippa-13b-qlora
9
+
10
+ This is a repository of my Llama-2-13b Qlora checkpoints of the [PIPPA-13b-ShareGPT](https://huggingface.co/datasets/kingbri/PIPPA-ShareGPT) dataset.
11
+
12
+ You can read more about the dataset on its relevant page. It's a ShareGPT reformat of the [PIPPA dataset](https://huggingface.co/datasets/PygmalionAI/PIPPA) by PygmalionAI. The reformat was done to allow for axolotl compatability.
13
+
14
+ ### Architecture
15
+
16
+ - **Model Architecture**: Llama-2-13b
17
+ - **Training Algorithm**: QLora
18
+ - **Dataset Used**: PIPPA-ShareGPT (pippa_sharegpt_trimmed.jsonl)
19
+
20
+ ### Training Details
21
+
22
+ - **Dataset**: [PIPPA-ShareGPT](https://huggingface.co/datasets/kingbri/PIPPA-ShareGPT)
23
+ - **Datset type**: ShareGPT
24
+ - **Training Parameters**: [See Here](https://gist.github.com/bdashore3/55ae04892f31609f2c3779c4a8a55408)
25
+ - **Training Environment**: Axolotl
26
+ - **sequence_len**: 4096
27
+
28
+ ## Notes
29
+
30
+ This Qlora was produced as an experiment to see how the public version of PIPPA can affect a model. As a result, I have no idea if this lora is of great quality or absolute garbage.
31
+
32
+ ## Acknowledgments
33
+
34
+ Thanks to:
35
+ - PygmalionAI: The creators of the PIPPA dataset
36
+ - Axolotl: Finetuning suite
37
+
38
+ ## Axolotl stuff
39
+
40
  ## Training procedure
41
 
42