KayO commited on
Commit
b747afb
·
verified ·
1 Parent(s): aef9cb3

End of training

Browse files
Files changed (5) hide show
  1. README.md +2 -2
  2. config.json +7 -1
  3. model.safetensors +1 -1
  4. tokenizer_config.json +1 -1
  5. training_args.bin +1 -1
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: KayO/SmolLM2-FT-MyDataset-WithChatTemplate
3
  library_name: transformers
4
  model_name: SmolLM2-FT-DPO
5
  tags:
@@ -13,7 +13,7 @@ licence: license
13
 
14
  # Model Card for SmolLM2-FT-DPO
15
 
16
- This model is a fine-tuned version of [KayO/SmolLM2-FT-MyDataset-WithChatTemplate](https://huggingface.co/KayO/SmolLM2-FT-MyDataset-WithChatTemplate).
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
 
1
  ---
2
+ base_model: HuggingFaceTB/SmolLM2-135M-Instruct
3
  library_name: transformers
4
  model_name: SmolLM2-FT-DPO
5
  tags:
 
13
 
14
  # Model Card for SmolLM2-FT-DPO
15
 
16
+ This model is a fine-tuned version of [HuggingFaceTB/SmolLM2-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct).
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "KayO/SmolLM2-FT-MyDataset-WithChatTemplate",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -27,6 +27,12 @@
27
  "rope_theta": 100000,
28
  "tie_word_embeddings": true,
29
  "torch_dtype": "float32",
 
 
 
 
 
 
30
  "transformers_version": "4.46.3",
31
  "use_cache": false,
32
  "vocab_size": 49152
 
1
  {
2
+ "_name_or_path": "HuggingFaceTB/SmolLM2-135M-Instruct",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
27
  "rope_theta": 100000,
28
  "tie_word_embeddings": true,
29
  "torch_dtype": "float32",
30
+ "transformers.js_config": {
31
+ "kv_cache_dtype": {
32
+ "fp16": "float16",
33
+ "q4f16": "float16"
34
+ }
35
+ },
36
  "transformers_version": "4.46.3",
37
  "use_cache": false,
38
  "vocab_size": 49152
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4064d60fd89a9ad885ea9ee44c9d72d4ee5b9bf47fd3008bb3f4c30adbcb8aab
3
  size 538090408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:313088167930aaa341d64283f97ecaf42e0d7e54af14d5677a8e185a08217b3f
3
  size 538090408
tokenizer_config.json CHANGED
@@ -143,7 +143,7 @@
143
  "<|im_end|>"
144
  ],
145
  "bos_token": "<|im_start|>",
146
- "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
147
  "clean_up_tokenization_spaces": false,
148
  "eos_token": "<|im_end|>",
149
  "model_max_length": 8192,
 
143
  "<|im_end|>"
144
  ],
145
  "bos_token": "<|im_start|>",
146
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful AI assistant named SmolLM, trained by Hugging Face<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
147
  "clean_up_tokenization_spaces": false,
148
  "eos_token": "<|im_end|>",
149
  "model_max_length": 8192,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4f17783bb256c0758604e2894a1875d3084251c561d55d0bb0498299d33aa1b
3
  size 6072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5a199b7ac2de43e8ca2400cfc7304bd9af3563738f45c44c7a92bfd58ef10b6
3
  size 6072