Upload LlamaForCausalLM
Browse files- README.md +3 -3
- generation_config.json +12 -2
README.md
CHANGED
@@ -1,12 +1,12 @@
|
|
1 |
---
|
2 |
-
|
|
|
3 |
datasets:
|
4 |
- ruggsea/stanford-encyclopedia-of-philosophy_chat_multi_turn
|
5 |
language:
|
6 |
- en
|
7 |
- it
|
8 |
-
|
9 |
-
- meta-llama/Llama-3.1-8B
|
10 |
---
|
11 |
# Llama3.1-SEP-Chat
|
12 |
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- meta-llama/Llama-3.1-8B
|
4 |
datasets:
|
5 |
- ruggsea/stanford-encyclopedia-of-philosophy_chat_multi_turn
|
6 |
language:
|
7 |
- en
|
8 |
- it
|
9 |
+
license: other
|
|
|
10 |
---
|
11 |
# Llama3.1-SEP-Chat
|
12 |
|
generation_config.json
CHANGED
@@ -2,9 +2,19 @@
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 128256,
|
4 |
"do_sample": true,
|
5 |
-
"eos_token_id":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
"pad_token_id": 128257,
|
7 |
-
"
|
|
|
8 |
"top_p": 0.9,
|
9 |
"transformers_version": "4.47.1"
|
10 |
}
|
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 128256,
|
4 |
"do_sample": true,
|
5 |
+
"eos_token_id": [
|
6 |
+
128257,
|
7 |
+
128001
|
8 |
+
],
|
9 |
+
"forced_eos_token_id": [
|
10 |
+
128257,
|
11 |
+
128001
|
12 |
+
],
|
13 |
+
"max_new_tokens": 512,
|
14 |
+
"no_repeat_ngram_size": 3,
|
15 |
"pad_token_id": 128257,
|
16 |
+
"repetition_penalty": 1.2,
|
17 |
+
"temperature": 0.7,
|
18 |
"top_p": 0.9,
|
19 |
"transformers_version": "4.47.1"
|
20 |
}
|