Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -7,6 +7,7 @@ import platform
|
|
7 |
|
8 |
model = ""
|
9 |
duration = 5
|
|
|
10 |
print(f"Is CUDA available: {torch.cuda.is_available()}")
|
11 |
print(f"CUDA device: {torch.cuda.get_device_name(torch.cuda.current_device())}")
|
12 |
print(f"CUDA version: {torch.version.cuda}")
|
@@ -16,8 +17,7 @@ print(f"Gradio version: {gr. __version__}")
|
|
16 |
|
17 |
|
18 |
"""
|
19 |
-
|
20 |
-
Packages that work::::::::::
|
21 |
Is CUDA available: True
|
22 |
CUDA device: NVIDIA A100-SXM4-80GB MIG 3g.40gb
|
23 |
CUDA version: 12.1
|
@@ -87,10 +87,6 @@ def respond(message, history: list[tuple[str, str]], model, system_message, max_
|
|
87 |
response += token
|
88 |
yield response
|
89 |
|
90 |
-
|
91 |
-
"""
|
92 |
-
For information on how to customize the ChatInterface, peruse the gradio docs: https://www.gradio.app/docs/chatinterface
|
93 |
-
"""
|
94 |
|
95 |
demo = gr.ChatInterface(
|
96 |
respond,
|
@@ -105,6 +101,5 @@ demo = gr.ChatInterface(
|
|
105 |
)
|
106 |
|
107 |
|
108 |
-
|
109 |
if __name__ == "__main__":
|
110 |
demo.launch()
|
|
|
7 |
|
8 |
model = ""
|
9 |
duration = 5
|
10 |
+
|
11 |
print(f"Is CUDA available: {torch.cuda.is_available()}")
|
12 |
print(f"CUDA device: {torch.cuda.get_device_name(torch.cuda.current_device())}")
|
13 |
print(f"CUDA version: {torch.version.cuda}")
|
|
|
17 |
|
18 |
|
19 |
"""
|
20 |
+
Packages ::::::::::
|
|
|
21 |
Is CUDA available: True
|
22 |
CUDA device: NVIDIA A100-SXM4-80GB MIG 3g.40gb
|
23 |
CUDA version: 12.1
|
|
|
87 |
response += token
|
88 |
yield response
|
89 |
|
|
|
|
|
|
|
|
|
90 |
|
91 |
demo = gr.ChatInterface(
|
92 |
respond,
|
|
|
101 |
)
|
102 |
|
103 |
|
|
|
104 |
if __name__ == "__main__":
|
105 |
demo.launch()
|