msho commited on
Commit
593ea27
·
verified ·
1 Parent(s): bf7aedf

Training in progress, step 33

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "gate_proj",
24
- "up_proj",
25
- "v_proj",
26
  "q_proj",
 
 
27
  "down_proj",
28
- "o_proj",
29
- "k_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "q_proj",
24
+ "v_proj",
25
+ "gate_proj",
26
  "down_proj",
27
+ "k_proj",
28
+ "up_proj",
29
+ "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83a9fd57c6db439461ce997290822484f3e9964dfa2c87266aeef24b272dce0b
3
  size 149969848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:caca8a2619bda1446233cda075831bdbd0ae0da0a12707ba71946b20fa5d2b84
3
  size 149969848
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 1024,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:529c972b6658bf07c40af6617d16336ec1125752cc14905f8d9357219aac0f1c
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86ee241536aa1604ab71face617c45e31aa09f34c4c493e1a93f53122fe56b48
3
  size 5432