Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -25,7 +25,7 @@ class MessageRequest(BaseModel):
|
|
| 25 |
message: str
|
| 26 |
language: str
|
| 27 |
|
| 28 |
-
repo_id = "
|
| 29 |
llm_client = InferenceClient(
|
| 30 |
model=repo_id,
|
| 31 |
token=os.getenv("HF_TOKEN"),
|
|
@@ -61,8 +61,8 @@ templates = Jinja2Templates(directory="static")
|
|
| 61 |
|
| 62 |
# Configure Llama index settings
|
| 63 |
Settings.llm = HuggingFaceInferenceAPI(
|
| 64 |
-
model_name="
|
| 65 |
-
tokenizer_name="
|
| 66 |
context_window=3000,
|
| 67 |
token=os.getenv("HF_TOKEN"),
|
| 68 |
max_new_tokens=512,
|
|
|
|
| 25 |
message: str
|
| 26 |
language: str
|
| 27 |
|
| 28 |
+
repo_id = "openai-community/gpt2"
|
| 29 |
llm_client = InferenceClient(
|
| 30 |
model=repo_id,
|
| 31 |
token=os.getenv("HF_TOKEN"),
|
|
|
|
| 61 |
|
| 62 |
# Configure Llama index settings
|
| 63 |
Settings.llm = HuggingFaceInferenceAPI(
|
| 64 |
+
model_name="openai-community/gpt2",
|
| 65 |
+
tokenizer_name="openai-community/gpt2",
|
| 66 |
context_window=3000,
|
| 67 |
token=os.getenv("HF_TOKEN"),
|
| 68 |
max_new_tokens=512,
|