CHZY-1 commited on
Commit
d09968e
·
verified ·
1 Parent(s): 25124c4

Re-trained QLora Adapter with 260 data (5 epoch)

Browse files
README.md CHANGED
@@ -56,5 +56,5 @@ The following hyperparameters were used during training:
56
  - PEFT 0.13.2
57
  - Transformers 4.44.2
58
  - Pytorch 2.4.1+cu121
59
- - Datasets 3.0.1
60
  - Tokenizers 0.19.1
 
56
  - PEFT 0.13.2
57
  - Transformers 4.44.2
58
  - Pytorch 2.4.1+cu121
59
+ - Datasets 3.0.2
60
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d19b6ab0d6078d854c0d387c113a79079379863d402e0ee61804ebb353168bb
3
  size 134235048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23dddfa490ab38917457b392acd668ad7630dc8194a7ce696c9565b4a56f76a3
3
  size 134235048
runs/Oct22_19-13-21_343480e751cb/events.out.tfevents.1729624405.343480e751cb.565.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1038e7f1c55efeecc645e947f946d4ec8a25b744644307d0fc32eecb0afd54d
3
+ size 8704
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a93b953fb61bdd7fcf217d3b3a53f13454a1030bfd4e5976d71f5004f1200688
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76c46029b8a10351b0ddfd8b1b0cab6a0b2f659431c4a3efe345b87f78ca9fb5
3
  size 5496