Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -10,17 +10,15 @@ import torch
|
|
10 |
# Use a pipeline as a high-level helper
|
11 |
from transformers import pipeline
|
12 |
|
|
|
13 |
messages = [
|
14 |
{"role": "user", "content": "Who are you?"},
|
15 |
]
|
16 |
-
|
17 |
-
model_name = "bmas10/DeepSeek-Llama-8-GGUF"
|
18 |
-
|
19 |
pipe = pipeline("text-generation", model=model_name)
|
20 |
pipe(messages)
|
21 |
|
22 |
|
23 |
-
|
24 |
def chat(input_text, history=[]):
|
25 |
history.append(input_text)
|
26 |
prompt = "\n".join(history) + "\nAI:" # Simple conversational format
|
@@ -82,3 +80,4 @@ demo = gr.ChatInterface(
|
|
82 |
|
83 |
if __name__ == "__main__":
|
84 |
demo.launch()
|
|
|
|
10 |
# Use a pipeline as a high-level helper
|
11 |
from transformers import pipeline
|
12 |
|
13 |
+
model_name = "bmas10/DeepSeek-Llama-8-GGUF"
|
14 |
messages = [
|
15 |
{"role": "user", "content": "Who are you?"},
|
16 |
]
|
|
|
|
|
|
|
17 |
pipe = pipeline("text-generation", model=model_name)
|
18 |
pipe(messages)
|
19 |
|
20 |
|
21 |
+
'''
|
22 |
def chat(input_text, history=[]):
|
23 |
history.append(input_text)
|
24 |
prompt = "\n".join(history) + "\nAI:" # Simple conversational format
|
|
|
80 |
|
81 |
if __name__ == "__main__":
|
82 |
demo.launch()
|
83 |
+
'''
|