darabos's picture
Use environment variables for selecting the models.
e8ea95d

LynxScribe in LynxKite

LynxKite UI for building LynxScribe chat applications. Also runs the chat application!

To run a chat UI for LynxScribe workspaces:

WEBUI_AUTH=false OPENAI_API_BASE_URL=http://localhost:8000/api/service/lynxkite_lynxscribe uvx open-webui serve

Or use Lynx WebUI instead of Open WebUI.

Run tests with:

uv run pytest

The LLM agent flow examples can use local models.

uv pip install infinity-emb[all]
infinity_emb v2 --model-id michaelfeil/bge-small-en-v1.5
uv pip install "sglang[all]>=0.4.2.post2" --find-links https://flashinfer.ai/whl/cu124/torch2.5/flashinfer/
python -m sglang.launch_server --model-path SultanR/SmolTulu-1.7b-Instruct --port 8080
export LLM_BASE_URL='http://localhost:8080/v1'
export LLM_MODEL='SultanR/SmolTulu-1.7b-Instruct'
export EMBEDDING_BASE_URL='http://localhost:7997/'
export EMBEDDING_MODEL='michaelfeil/bge-small-en-v1.5'
lynxkite