Update README.md
Browse files
README.md
CHANGED
@@ -45,6 +45,18 @@ USER: prompt
|
|
45 |
ASSISTANT:
|
46 |
```
|
47 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
48 |
## THE FILES IN MAIN BRANCH REQUIRES LATEST LLAMA.CPP (May 19th 2023 - commit 2d5db48)!
|
49 |
|
50 |
llama.cpp recently made another breaking change to its quantisation methods - https://github.com/ggerganov/llama.cpp/pull/1508
|
|
|
45 |
ASSISTANT:
|
46 |
```
|
47 |
|
48 |
+
## Context length with GGML
|
49 |
+
|
50 |
+
The base Airoboros GPT4 models have an increased context length of 4096.
|
51 |
+
|
52 |
+
However this GGML conversion appears to still have the default 2048 context.
|
53 |
+
|
54 |
+
I have experimented with llama.cpp's `-n 4096` parameter to specify a context of 4096 but it so far always results in gibberish output.
|
55 |
+
|
56 |
+
I will investigate this further and upload a correct model if this proves necessary.
|
57 |
+
|
58 |
+
For now, please assume this GGML to have a context of 2048.
|
59 |
+
|
60 |
## THE FILES IN MAIN BRANCH REQUIRES LATEST LLAMA.CPP (May 19th 2023 - commit 2d5db48)!
|
61 |
|
62 |
llama.cpp recently made another breaking change to its quantisation methods - https://github.com/ggerganov/llama.cpp/pull/1508
|