|
[generator] |
|
PROVIDER = huggingface |
|
MODEL = meta-llama/Meta-Llama-3-8B-Instruct |
|
MAX_TOKENS = 768 |
|
TEMPERATURE = 0.2 |
|
INFERENCE_PROVIDER = novita |
|
ORGANIZATION = GIZ |
|
|
|
[reader] |
|
TYPE = INF_PROVIDERS |
|
INF_PROVIDER_MODEL = meta-llama/Llama-3.1-8B-Instruct |
|
DEDICATED_MODEL = meta-llama/Llama-3.1-8B-Instruct |
|
DEDICATED_ENDPOINT = https://qu2d8m6dmsollhly.us-east-1.aws.endpoints.huggingface.cloud |
|
NVIDIA_MODEL = meta-llama/Llama-3.1-8B-Instruct |
|
NVIDIA_ENDPOINT = https://huggingface.co/api/integrations/dgx/v1 |
|
MAX_TOKENS = 768 |
|
INF_PROVIDER = nebius |
|
|
|
[app] |
|
dropdown_default = Annual Consolidated OAG 2024 |