Update README.md
Browse files
README.md
CHANGED
@@ -85,30 +85,38 @@ into a single file, using the same order.
|
|
85 |
|
86 |
## Prompting
|
87 |
|
88 |
-
|
89 |
|
90 |
```
|
91 |
-
<BOS_TOKEN>
|
92 |
<|START_OF_TURN_TOKEN|>
|
93 |
-
<|USER_TOKEN|>
|
94 |
<|START_OF_TURN_TOKEN|>
|
95 |
-
<|CHATBOT_TOKEN|>
|
96 |
```
|
97 |
|
98 |
-
|
99 |
|
100 |
```
|
101 |
-
|
|
|
|
|
102 |
<|START_OF_TURN_TOKEN|>
|
103 |
-
<|
|
|
|
|
|
|
|
|
|
|
|
104 |
<|START_OF_TURN_TOKEN|>
|
105 |
-
<|
|
106 |
```
|
107 |
|
108 |
The maximum context size of this model is 8192 tokens. These llamafiles
|
109 |
use a default context size of 512 tokens. Whenever you need the maximum
|
110 |
context size to be available with llamafile for any given model, you can
|
111 |
-
pass the `-c 0` flag.
|
|
|
112 |
|
113 |
## License
|
114 |
|
|
|
85 |
|
86 |
## Prompting
|
87 |
|
88 |
+
Command-line instruction template:
|
89 |
|
90 |
```
|
91 |
+
./aya-23-8B.Q8_0.llamafile --log-disable --silent-prompt -p '<BOS_TOKEN>
|
92 |
<|START_OF_TURN_TOKEN|>
|
93 |
+
<|USER_TOKEN|>Who is the president?<|END_OF_TURN_TOKEN|>
|
94 |
<|START_OF_TURN_TOKEN|>
|
95 |
+
<|CHATBOT_TOKEN|>'
|
96 |
```
|
97 |
|
98 |
+
"Prompt Template" for web GUI interface:
|
99 |
|
100 |
```
|
101 |
+
<BOS_TOKEN>
|
102 |
+
<|SYSTEM_TOKEN|>{{prompt}}<|END_OF_TURN_TOKEN|>
|
103 |
+
{{history}}
|
104 |
<|START_OF_TURN_TOKEN|>
|
105 |
+
<|CHATBOT_TOKEN|>
|
106 |
+
```
|
107 |
+
|
108 |
+
"Chat history template" for web GUI interface:
|
109 |
+
|
110 |
+
```
|
111 |
<|START_OF_TURN_TOKEN|>
|
112 |
+
<|USER_TOKEN|>{{message}}<|END_OF_TURN_TOKEN|>
|
113 |
```
|
114 |
|
115 |
The maximum context size of this model is 8192 tokens. These llamafiles
|
116 |
use a default context size of 512 tokens. Whenever you need the maximum
|
117 |
context size to be available with llamafile for any given model, you can
|
118 |
+
pass the `-c 0` flag. The temperature on these llamafiles is set to zero
|
119 |
+
by default, because it helps. This can be changed, e.g. `--temp 0.8`.
|
120 |
|
121 |
## License
|
122 |
|