|
--- |
|
license: apache-2.0 |
|
--- |
|
This is a llamafile for [BioMistral-7B-DARE](https://huggingface.co/BioMistral/BioMistral-7B-DARE) quantized to q5-k-m. |
|
|
|
Safetensors came directly from their [source](https://huggingface.co/Nexusflow/Starling-LM-7B-beta/tree/main), |
|
converted to gguf and then quantized to q5-k-m with [llama.cpp](https://github.com/ggerganov/llama.cpp), |
|
and then zipped into a llamafile using [Mozilla's awesome project](https://github.com/Mozilla-Ocho/llamafile). |
|
|
|
|
|
|
|
-= Llamafile =- |
|
|
|
Llamafiles are a standalone executable that run an LLM server locally on a variety of operating systems including FreeBSD, Windows, Windows via WSL, Linux, and Mac. |
|
The same file works everywhere, I've tested several of these on FreeBSD, Windows, Windows via WSL, and Linux. |
|
You just download the .llamafile, (chmod +x or rename to .exe as needed), run it, open the chat interface in a browser, and interact. |
|
Options can be passed in to expose the api etc. See their [docs](https://github.com/Mozilla-Ocho/llamafile) for details. |
|
|
|
[Mozilla Blog Announcement for Llamafile](https://hacks.mozilla.org/2023/11/introducing-llamafile/) |
|
|
|
|
|
- Windows note: If it's over 4gb and you want to use it on Windows, you'll have to run it from WSL. |
|
|
|
- WSL note: If you get the error about APE, and the recommended command |
|
|
|
`sudo sh -c 'echo -1 > /proc/sys/fs/binfmt_misc/WSLInterop'` |
|
|
|
doesn't work, the WSLInterop file might be named something else. I had success with |
|
|
|
`sudo sh -c 'echo -1 > /proc/sys/fs/binfmt_misc/WSLInterop-late'` |
|
|
|
If that fails too, just navigate to `/proc/sys/fs/binfmt_msc` and see what files look like `WSLInterop` |
|
and echo a -1 to whatever they're called by changing that part of the recommended command. |
|
|
|
|
|
- FreeBSD note: Yes, it actually works on a fresh install of FreeBSD. |