msho commited on
Commit
bf7aedf
·
verified ·
1 Parent(s): f08a4f3

Training in progress, step 33

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "down_proj",
24
- "k_proj",
25
  "gate_proj",
 
26
  "v_proj",
 
 
27
  "o_proj",
28
- "up_proj",
29
- "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "gate_proj",
24
+ "up_proj",
25
  "v_proj",
26
+ "q_proj",
27
+ "down_proj",
28
  "o_proj",
29
+ "k_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20e5c5547ef6726c3234ea9cc7e93078d37fb91fc6ab6282714a319759991c3f
3
  size 149969848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83a9fd57c6db439461ce997290822484f3e9964dfa2c87266aeef24b272dce0b
3
  size 149969848
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cea0109244c18a333ae9f5320a43137a21c222e83a59c9631875c124aef90f69
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:529c972b6658bf07c40af6617d16336ec1125752cc14905f8d9357219aac0f1c
3
  size 5432