Remove .env.template file
Browse files- .env.template +0 -46
.env.template
DELETED
@@ -1,46 +0,0 @@
|
|
1 |
-
# Required: Hugging Face token with read/write permissions for repositories and inference API
|
2 |
-
# Get it from: https://huggingface.co/settings/tokens
|
3 |
-
HF_TOKEN=your_hugging_face_token_here
|
4 |
-
|
5 |
-
# Model Configuration
|
6 |
-
# Choose ONE of the following model setups:
|
7 |
-
|
8 |
-
## 1. For Hugging Face Serverless Inference (Default):
|
9 |
-
MODEL=meta-llama/Llama-3.1-8B-Instruct
|
10 |
-
# MODEL=Qwen/Qwen2.5-1.5B-Instruct
|
11 |
-
# MODEL=mistralai/Mixtral-8x7B-Instruct-v0.1
|
12 |
-
# MODEL=HuggingFaceH4/zephyr-7b-beta
|
13 |
-
|
14 |
-
## 2. For OpenAI-compatible API:
|
15 |
-
# OPENAI_BASE_URL=https://api.openai.com/v1/
|
16 |
-
# MODEL=gpt-4
|
17 |
-
# API_KEY=your_openai_api_key_here
|
18 |
-
|
19 |
-
## 3. For Ollama:
|
20 |
-
# OLLAMA_BASE_URL=http://127.0.0.1:11434/
|
21 |
-
# MODEL=qwen2.5:32b-instruct-q5_K_S
|
22 |
-
# TOKENIZER_ID=Qwen/Qwen2.5-32B-Instruct
|
23 |
-
|
24 |
-
## 4. For VLLM:
|
25 |
-
# VLLM_BASE_URL=http://127.0.0.1:8000/
|
26 |
-
# MODEL=Qwen/Qwen2.5-1.5B-Instruct
|
27 |
-
# TOKENIZER_ID=Qwen/Qwen2.5-1.5B-Instruct
|
28 |
-
|
29 |
-
## 5. For Hugging Face Inference Endpoints or TGI:
|
30 |
-
# HUGGINGFACE_BASE_URL=http://127.0.0.1:3000/
|
31 |
-
# TOKENIZER_ID=meta-llama/Llama-3.1-8B-Instruct
|
32 |
-
|
33 |
-
# Generation Settings (Optional - these are the defaults)
|
34 |
-
MAX_NUM_TOKENS=2048
|
35 |
-
MAX_NUM_ROWS=1000
|
36 |
-
DEFAULT_BATCH_SIZE=5
|
37 |
-
|
38 |
-
# Chat/SFT Configuration
|
39 |
-
# Required for chat data generation with Llama or Qwen models
|
40 |
-
# Options: "llama3", "qwen2", or custom template string
|
41 |
-
MAGPIE_PRE_QUERY_TEMPLATE=llama3
|
42 |
-
|
43 |
-
# Optional: Argilla Integration
|
44 |
-
# Follow https://docs.argilla.io/latest/getting_started/quickstart/
|
45 |
-
# ARGILLA_API_URL=https://[your-owner-name]-[your_space_name].hf.space
|
46 |
-
# ARGILLA_API_KEY=your_argilla_api_key_here
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|