Spaces:
Running
Running
# LynxScribe in LynxKite | |
LynxKite UI for building LynxScribe chat applications. Also runs the chat application! | |
To run a chat UI for LynxScribe workspaces: | |
```bash | |
WEBUI_AUTH=false OPENAI_API_BASE_URL=http://localhost:8000/api/service/lynxkite_lynxscribe uvx open-webui serve | |
``` | |
Or use [Lynx WebUI](https://github.com/biggraph/lynx-webui/) instead of Open WebUI. | |
Run tests with: | |
```bash | |
uv run pytest | |
``` | |
The LLM agent flow examples can use local models. | |
```bash | |
uv pip install infinity-emb[all] | |
infinity_emb v2 --model-id michaelfeil/bge-small-en-v1.5 | |
uv pip install "sglang[all]>=0.4.2.post2" --find-links https://flashinfer.ai/whl/cu124/torch2.5/flashinfer/ | |
python -m sglang.launch_server --model-path SultanR/SmolTulu-1.7b-Instruct --port 8080 | |
export LLM_BASE_URL='http://localhost:8080/v1' | |
export LLM_MODEL='SultanR/SmolTulu-1.7b-Instruct' | |
export EMBEDDING_BASE_URL='http://localhost:7997/' | |
export EMBEDDING_MODEL='michaelfeil/bge-small-en-v1.5' | |
lynxkite | |
``` | |