cwiz commited on
Commit
f22d907
·
1 Parent(s): 8ed3053

Upload 5 files

Browse files
adapter_config.json CHANGED
@@ -4,11 +4,14 @@
4
  "fan_in_fan_out": false,
5
  "inference_mode": true,
6
  "init_lora_weights": true,
 
 
7
  "lora_alpha": 64,
8
  "lora_dropout": 0.05,
9
  "modules_to_save": null,
10
  "peft_type": "LORA",
11
- "r": 128,
 
12
  "target_modules": [
13
  "q_proj",
14
  "v_proj"
 
4
  "fan_in_fan_out": false,
5
  "inference_mode": true,
6
  "init_lora_weights": true,
7
+ "layers_pattern": null,
8
+ "layers_to_transform": null,
9
  "lora_alpha": 64,
10
  "lora_dropout": 0.05,
11
  "modules_to_save": null,
12
  "peft_type": "LORA",
13
+ "r": 32,
14
+ "revision": null,
15
  "target_modules": [
16
  "q_proj",
17
  "v_proj"
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ea43d609d3eb460fefb36d5d2fccb7d4c555cd3d01188b9e2c136b3495e931d
3
- size 268481485
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55647e7ad088def6a89f90d0f5d5c2b06443df354213bc603ecb61c19b40b096
3
+ size 67154893
training_log.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name": "decapoda-research_llama-7b-hf",
3
+ "base_model_class": "LlamaForCausalLM",
4
+ "base_loaded_in_4bit": false,
5
+ "base_loaded_in_8bit": true,
6
+ "loss": 2.0215,
7
+ "learning_rate": 0.0,
8
+ "epoch": 3.0,
9
+ "current_steps": 57599,
10
+ "train_runtime": 58209.0858,
11
+ "train_samples_per_second": 3.961,
12
+ "train_steps_per_second": 0.031,
13
+ "total_flos": 2.344218918068945e+18,
14
+ "train_loss": 2.1099784564971924
15
+ }
training_parameters.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lora_name": "schizo-lm",
3
+ "always_override": true,
4
+ "save_steps": 0.0,
5
+ "micro_batch_size": 4,
6
+ "batch_size": 128,
7
+ "epochs": 3.0,
8
+ "learning_rate": "3e-4",
9
+ "lr_scheduler_type": "linear",
10
+ "lora_rank": 32,
11
+ "lora_alpha": 64,
12
+ "lora_dropout": 0.05,
13
+ "cutoff_len": 256,
14
+ "dataset": "None",
15
+ "eval_dataset": "None",
16
+ "format": "None",
17
+ "eval_steps": 100.0,
18
+ "raw_text_file": "comb4",
19
+ "overlap_len": 128,
20
+ "newline_favor_len": 128,
21
+ "higher_rank_limit": false,
22
+ "warmup_steps": 100.0,
23
+ "optimizer": "adamw_torch",
24
+ "hard_cut_string": "\\n\\n\\n",
25
+ "train_only_after": "",
26
+ "stop_at_loss": 0
27
+ }
training_prompt.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "template_type": "raw_text"
3
+ }