Spaces:
Sleeping
Sleeping
kwabs22
commited on
Commit
·
a7e3cb8
1
Parent(s):
1c670bf
CUDA location is probably zero issue
Browse files
app.py
CHANGED
@@ -11,7 +11,7 @@ model = None
|
|
11 |
def loadmodel():
|
12 |
global tokenizer, model
|
13 |
tokenizer = AutoTokenizer.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16")
|
14 |
-
model = AutoModelForCausalLM.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16", torch_dtype=
|
15 |
#model = model.to('cuda') # Move the model to GPU if available
|
16 |
pass
|
17 |
|
|
|
11 |
def loadmodel():
|
12 |
global tokenizer, model
|
13 |
tokenizer = AutoTokenizer.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16")
|
14 |
+
model = AutoModelForCausalLM.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16", torch_dtype='auto', device_map= 'auto') #torch_dtype=torch.float16
|
15 |
#model = model.to('cuda') # Move the model to GPU if available
|
16 |
pass
|
17 |
|