darabos's picture
Use environment variables for selecting the models.
e8ea95d
# LynxScribe in LynxKite
LynxKite UI for building LynxScribe chat applications. Also runs the chat application!
To run a chat UI for LynxScribe workspaces:
```bash
WEBUI_AUTH=false OPENAI_API_BASE_URL=http://localhost:8000/api/service/lynxkite_lynxscribe uvx open-webui serve
```
Or use [Lynx WebUI](https://github.com/biggraph/lynx-webui/) instead of Open WebUI.
Run tests with:
```bash
uv run pytest
```
The LLM agent flow examples can use local models.
```bash
uv pip install infinity-emb[all]
infinity_emb v2 --model-id michaelfeil/bge-small-en-v1.5
uv pip install "sglang[all]>=0.4.2.post2" --find-links https://flashinfer.ai/whl/cu124/torch2.5/flashinfer/
python -m sglang.launch_server --model-path SultanR/SmolTulu-1.7b-Instruct --port 8080
export LLM_BASE_URL='http://localhost:8080/v1'
export LLM_MODEL='SultanR/SmolTulu-1.7b-Instruct'
export EMBEDDING_BASE_URL='http://localhost:7997/'
export EMBEDDING_MODEL='michaelfeil/bge-small-en-v1.5'
lynxkite
```