treasure4l commited on
Commit
d5f406a
·
verified ·
1 Parent(s): 942c373

End of training

Browse files
README.md CHANGED
@@ -4,9 +4,9 @@ library_name: transformers
4
  model_name: Mistral_7B-DPO
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - dpo
9
- - unsloth
10
  licence: license
11
  ---
12
 
@@ -35,11 +35,11 @@ This model was trained with DPO, a method introduced in [Direct Preference Optim
35
 
36
  ### Framework versions
37
 
38
- - TRL: 0.13.0
39
- - Transformers: 4.49.0.dev0
40
- - Pytorch: 2.1.0+cu118
41
- - Datasets: 3.2.0
42
- - Tokenizers: 0.21.0
43
 
44
  ## Citations
45
 
 
4
  model_name: Mistral_7B-DPO
5
  tags:
6
  - generated_from_trainer
7
+ - unsloth
8
  - trl
9
  - dpo
 
10
  licence: license
11
  ---
12
 
 
35
 
36
  ### Framework versions
37
 
38
+ - TRL: 0.15.2
39
+ - Transformers: 4.50.0
40
+ - Pytorch: 2.6.0
41
+ - Datasets: 3.4.1
42
+ - Tokenizers: 0.21.1
43
 
44
  ## Citations
45
 
adapter_config.json CHANGED
@@ -3,6 +3,7 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
5
  "bias": "none",
 
6
  "eva_config": null,
7
  "exclude_modules": null,
8
  "fan_in_fan_out": false,
@@ -23,15 +24,16 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "v_proj",
27
  "q_proj",
28
- "k_proj",
29
- "gate_proj",
30
- "o_proj",
31
  "up_proj",
32
- "down_proj"
 
 
33
  ],
34
  "task_type": "CAUSAL_LM",
 
35
  "use_dora": false,
36
  "use_rslora": false
37
  }
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
5
  "bias": "none",
6
+ "corda_config": null,
7
  "eva_config": null,
8
  "exclude_modules": null,
9
  "fan_in_fan_out": false,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "gate_proj",
28
  "v_proj",
29
  "q_proj",
 
 
 
30
  "up_proj",
31
+ "down_proj",
32
+ "k_proj",
33
+ "o_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
+ "trainable_token_indices": null,
37
  "use_dora": false,
38
  "use_rslora": false
39
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3216c79cc3f3dcb792e9483861df1582229f938e12a6c3057a49f7a377e49037
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ccb9d1fedd46f63f17f184aec0d805358f5aea908520004997fc0497934028d
3
  size 167832240
tokenizer_config.json CHANGED
@@ -6173,7 +6173,6 @@
6173
  }
6174
  },
6175
  "bos_token": "<s>",
6176
- "chat_template": "alpaca",
6177
  "clean_up_tokenization_spaces": false,
6178
  "eos_token": "</s>",
6179
  "extra_special_tokens": {},
 
6173
  }
6174
  },
6175
  "bos_token": "<s>",
 
6176
  "clean_up_tokenization_spaces": false,
6177
  "eos_token": "</s>",
6178
  "extra_special_tokens": {},
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0700e29a23833f5033edc6cd73e17db0f62f06c4f2f3fa08d131bde0b889b66
3
- size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2382abf2eb9bf526959044d2e4372aeff604be2da83dd452144fd02d5ad02af3
3
+ size 6264