Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -2,16 +2,6 @@ import os
|
|
2 |
from groq import Groq
|
3 |
import gradio as gr
|
4 |
from transformers import AutoModel, AutoConfig
|
5 |
-
|
6 |
-
hf_token = os.environ.get("HF_TOKEN")
|
7 |
-
|
8 |
-
model_repo = "HusseinEid/lora_model"
|
9 |
-
config_file = "config.json"
|
10 |
-
|
11 |
-
# Load the configuration directly from the repository
|
12 |
-
config = AutoConfig.from_pretrained(model_repo, config_file_name=config_file, token=hf_token)
|
13 |
-
# Load the model from Hugging Face using the config
|
14 |
-
model = AutoModel.from_pretrained(model_repo, config=config, token=hf_token)
|
15 |
|
16 |
client = Groq(api_key = os.environ.get("GROQ_API_KEY"), )
|
17 |
|
@@ -35,7 +25,7 @@ async def chat_groq(message, history):
|
|
35 |
response_content = ''
|
36 |
|
37 |
stream = client. chat.completions.create(
|
38 |
-
model=
|
39 |
messages=messages,
|
40 |
max_tokens=1024,
|
41 |
temperature=1.2,
|
|
|
2 |
from groq import Groq
|
3 |
import gradio as gr
|
4 |
from transformers import AutoModel, AutoConfig
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
|
6 |
client = Groq(api_key = os.environ.get("GROQ_API_KEY"), )
|
7 |
|
|
|
25 |
response_content = ''
|
26 |
|
27 |
stream = client. chat.completions.create(
|
28 |
+
model="llama3-8b-8192",
|
29 |
messages=messages,
|
30 |
max_tokens=1024,
|
31 |
temperature=1.2,
|