Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -12,18 +12,22 @@ from huggingface_hub import snapshot_download
|
|
| 12 |
from vptq.app_utils import get_chat_loop_generator
|
| 13 |
|
| 14 |
models = [
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
{
|
| 16 |
"name": "VPTQ-community/Meta-Llama-3.1-8B-Instruct-v8-k65536-256-woft",
|
| 17 |
"bits": "3 bits"
|
| 18 |
},
|
| 19 |
{
|
| 20 |
-
"name": "VPTQ-community/Meta-Llama-3.1-
|
| 21 |
-
"bits": "
|
| 22 |
},
|
| 23 |
{
|
| 24 |
-
"name": "VPTQ-community/
|
| 25 |
-
"bits": "
|
| 26 |
-
}
|
| 27 |
]
|
| 28 |
|
| 29 |
# Queues for storing historical data (saving the last 100 GPU utilization and memory usage values)
|
|
|
|
| 12 |
from vptq.app_utils import get_chat_loop_generator
|
| 13 |
|
| 14 |
models = [
|
| 15 |
+
{
|
| 16 |
+
"name": "VPTQ-community/Meta-Llama-3.1-8B-Instruct-v12-k65536-4096-woft",
|
| 17 |
+
"bits": "2.3 bits"
|
| 18 |
+
},
|
| 19 |
{
|
| 20 |
"name": "VPTQ-community/Meta-Llama-3.1-8B-Instruct-v8-k65536-256-woft",
|
| 21 |
"bits": "3 bits"
|
| 22 |
},
|
| 23 |
{
|
| 24 |
+
"name": "VPTQ-community/Meta-Llama-3.1-8B-Instruct-v8-k65536-4096-woft",
|
| 25 |
+
"bits": "3.5 bits"
|
| 26 |
},
|
| 27 |
{
|
| 28 |
+
"name": "VPTQ-community/Meta-Llama-3.1-70B-Instruct-v8-k32768-0-woft",
|
| 29 |
+
"bits": "1.85 bits"
|
| 30 |
+
},
|
| 31 |
]
|
| 32 |
|
| 33 |
# Queues for storing historical data (saving the last 100 GPU utilization and memory usage values)
|