imdatta0 commited on
Commit
d6d7d68
·
verified ·
1 Parent(s): e7c5715

Training in progress, step 4

Browse files
adapter_config.json CHANGED
@@ -51,7 +51,7 @@
51
  40
52
  ],
53
  "loftq_config": {},
54
- "lora_alpha": 8,
55
  "lora_dropout": 0,
56
  "megatron_config": null,
57
  "megatron_core": "megatron.core",
@@ -61,13 +61,13 @@
61
  "rank_pattern": {},
62
  "revision": "unsloth",
63
  "target_modules": [
64
- "up_proj",
65
  "k_proj",
 
 
 
66
  "v_proj",
67
  "o_proj",
68
- "down_proj",
69
- "q_proj",
70
- "gate_proj"
71
  ],
72
  "task_type": "CAUSAL_LM"
73
  }
 
51
  40
52
  ],
53
  "loftq_config": {},
54
+ "lora_alpha": 16,
55
  "lora_dropout": 0,
56
  "megatron_config": null,
57
  "megatron_core": "megatron.core",
 
61
  "rank_pattern": {},
62
  "revision": "unsloth",
63
  "target_modules": [
 
64
  "k_proj",
65
+ "gate_proj",
66
+ "q_proj",
67
+ "down_proj",
68
  "v_proj",
69
  "o_proj",
70
+ "up_proj"
 
 
71
  ],
72
  "task_type": "CAUSAL_LM"
73
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0bdeb961475868c1973b03b3e8b17456bda247927413e45725c6bd9f470749a2
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19f610a67a7393050fb2cb9173f4560a677011dea770c2485195466419f97242
3
  size 83945296
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d33a4ef2bbc9ec7635f7f53470603d89aa1b3036909d2bca8bcaaa5dc675c12
3
- size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:948783678db3f4896c166743a7e27686a338bee6ae0b474c2efc89e57833e099
3
+ size 5176