jonaschua commited on
Commit
da3c5f3
·
verified ·
1 Parent(s): 6e03532

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -7
app.py CHANGED
@@ -7,6 +7,7 @@ import platform
7
 
8
  model = ""
9
  duration = 5
 
10
  print(f"Is CUDA available: {torch.cuda.is_available()}")
11
  print(f"CUDA device: {torch.cuda.get_device_name(torch.cuda.current_device())}")
12
  print(f"CUDA version: {torch.version.cuda}")
@@ -16,8 +17,7 @@ print(f"Gradio version: {gr. __version__}")
16
 
17
 
18
  """
19
- For more information on `huggingface_hub` Inference API support, please check the docs: https://huggingface.co/docs/huggingface_hub/v0.22.2/en/guides/inference
20
- Packages that work::::::::::
21
  Is CUDA available: True
22
  CUDA device: NVIDIA A100-SXM4-80GB MIG 3g.40gb
23
  CUDA version: 12.1
@@ -87,10 +87,6 @@ def respond(message, history: list[tuple[str, str]], model, system_message, max_
87
  response += token
88
  yield response
89
 
90
-
91
- """
92
- For information on how to customize the ChatInterface, peruse the gradio docs: https://www.gradio.app/docs/chatinterface
93
- """
94
 
95
  demo = gr.ChatInterface(
96
  respond,
@@ -105,6 +101,5 @@ demo = gr.ChatInterface(
105
  )
106
 
107
 
108
-
109
  if __name__ == "__main__":
110
  demo.launch()
 
7
 
8
  model = ""
9
  duration = 5
10
+
11
  print(f"Is CUDA available: {torch.cuda.is_available()}")
12
  print(f"CUDA device: {torch.cuda.get_device_name(torch.cuda.current_device())}")
13
  print(f"CUDA version: {torch.version.cuda}")
 
17
 
18
 
19
  """
20
+ Packages ::::::::::
 
21
  Is CUDA available: True
22
  CUDA device: NVIDIA A100-SXM4-80GB MIG 3g.40gb
23
  CUDA version: 12.1
 
87
  response += token
88
  yield response
89
 
 
 
 
 
90
 
91
  demo = gr.ChatInterface(
92
  respond,
 
101
  )
102
 
103
 
 
104
  if __name__ == "__main__":
105
  demo.launch()