kwabs22 commited on
Commit
a7e3cb8
·
1 Parent(s): 1c670bf

CUDA location is probably zero issue

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -11,7 +11,7 @@ model = None
11
  def loadmodel():
12
  global tokenizer, model
13
  tokenizer = AutoTokenizer.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16")
14
- model = AutoModelForCausalLM.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16", torch_dtype=torch.float16, device_map= 'auto')
15
  #model = model.to('cuda') # Move the model to GPU if available
16
  pass
17
 
 
11
  def loadmodel():
12
  global tokenizer, model
13
  tokenizer = AutoTokenizer.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16")
14
+ model = AutoModelForCausalLM.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16", torch_dtype='auto', device_map= 'auto') #torch_dtype=torch.float16
15
  #model = model.to('cuda') # Move the model to GPU if available
16
  pass
17