Commit
·
9f6103d
1
Parent(s):
9304d02
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
datasets:
|
3 |
+
- lemonilia/LimaRP
|
4 |
+
language:
|
5 |
+
- en
|
6 |
+
tags:
|
7 |
+
- llama
|
8 |
+
- llama 2
|
9 |
+
- smol_llama
|
10 |
+
---
|
11 |
+
# smol_llama-220M-GQA-32k-theta-sft-limarp
|
12 |
+
|
13 |
+
Experimental model meant to serve as a long-context speculative decoding model. This one is specifically for models trained on the LimaRP prompt format.
|
14 |
+
|
15 |
+
Created using [Doctor-Shotgun/smol_llama-220M-GQA-32k-theta-sft](https://huggingface.co/Doctor-Shotgun/smol_llama-220M-GQA-32k-theta-sft) and finetuning at 32768 context length on several instruction datasets.
|
16 |
+
|
17 |
+
This variant uses the rope theta (rope frequency base) method for context extension.
|
18 |
+
|
19 |
+
The trained instruction format is LimaRP Alpaca:
|
20 |
+
```
|
21 |
+
### Instruction:
|
22 |
+
Character's Persona: {bot character description}
|
23 |
+
|
24 |
+
User's Persona: {user character description}
|
25 |
+
|
26 |
+
Scenario: {what happens in the story}
|
27 |
+
|
28 |
+
Play the role of Character. Taking the above information into consideration, you must engage in a roleplaying chat with User below this line. Do not write dialogues and narration for User.
|
29 |
+
|
30 |
+
### Input:
|
31 |
+
User: {utterance}
|
32 |
+
|
33 |
+
### Response:
|
34 |
+
Character: {utterance}
|
35 |
+
|
36 |
+
### Input
|
37 |
+
User: {utterance}
|
38 |
+
|
39 |
+
### Response:
|
40 |
+
Character: {utterance}
|
41 |
+
|
42 |
+
(etc.)
|
43 |
+
```
|