gobean commited on
Commit
a7f2521
·
verified ·
1 Parent(s): 81d6ecc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -0
README.md CHANGED
@@ -1,3 +1,25 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+ This is a llamafile for [BioMistral-7B-DARE](https://huggingface.co/BioMistral/BioMistral-7B-DARE) quantized to q5-k-m.
5
+
6
+ Safetensors came directly from their [source](https://huggingface.co/Nexusflow/Starling-LM-7B-beta/tree/main),
7
+ converted to gguf and then quantized to q5-k-m with [llama.cpp](https://github.com/ggerganov/llama.cpp),
8
+ and then zipped into a llamafile using [Mozilla's awesome project](https://github.com/Mozilla-Ocho/llamafile).
9
+
10
+ It's over 4gb so if you want to use it on Windows you'll have to run it from WSL.
11
+
12
+ WSL note: If you get the error about APE, and the recommended command
13
+
14
+ `sudo sh -c 'echo -1 > /proc/sys/fs/binfmt_misc/WSLInterop'`
15
+
16
+ doesn't work, the file might be named something else so I had success with
17
+
18
+ `sudo sh -c 'echo -1 > /proc/sys/fs/binfmt_misc/WSLInterop-late'`
19
+
20
+ If that fails too, just navigate to `/proc/sys/fs/binfmt_msc` and see what files look like `WSLInterop` and echo a -1 to whatever it's called by changing that part of the recommended command.
21
+
22
+
23
+ Llamafiles are a standalone executable that run an LLM server locally on a variety of operating systems.
24
+ You just run it, open the chat interface in a browser, and interact.
25
+ Options can be passed in to expose the api etc.