Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -21,17 +21,17 @@ access_token = os.environ["TOKEN"]
|
|
21 |
model = "NousResearch/Llama-2-7b-hf"
|
22 |
peft_model = "FinGPT/fingpt-forecaster_sz50_llama2-7B_lora"
|
23 |
|
24 |
-
tokenizer = AutoTokenizer.from_pretrained(model, token = access_token, device_map="
|
25 |
tokenizer.pad_token = tokenizer.eos_token
|
26 |
tokenizer.padding_side = "right"
|
27 |
|
28 |
streamer = TextStreamer(tokenizer)
|
29 |
|
30 |
-
model = AutoModelForCausalLM.from_pretrained(model, trust_remote_code=True, token = access_token, device_map="
|
31 |
-
model = PeftModel.from_pretrained(model, peft_model, device_map="
|
32 |
|
33 |
-
|
34 |
-
device = 'cuda:0'
|
35 |
model = model.to(device)
|
36 |
model = model.eval()
|
37 |
|
@@ -401,9 +401,9 @@ def ask(symbol, weeks_before, withbasic):
|
|
401 |
print(model.device)
|
402 |
|
403 |
inputs = tokenizer(pt, return_tensors='pt')
|
|
|
404 |
#inputs = {key: value.to(model.device) for key, value in inputs.items()}
|
405 |
-
#inputs = {key: value.to(
|
406 |
-
inputs = {key: value.to('cuda:0') for key, value in inputs.items()}
|
407 |
print("Inputs loaded onto devices.")
|
408 |
|
409 |
res = model.generate(
|
|
|
21 |
model = "NousResearch/Llama-2-7b-hf"
|
22 |
peft_model = "FinGPT/fingpt-forecaster_sz50_llama2-7B_lora"
|
23 |
|
24 |
+
tokenizer = AutoTokenizer.from_pretrained(model, token = access_token, device_map="auto", trust_remote_code=True)
|
25 |
tokenizer.pad_token = tokenizer.eos_token
|
26 |
tokenizer.padding_side = "right"
|
27 |
|
28 |
streamer = TextStreamer(tokenizer)
|
29 |
|
30 |
+
model = AutoModelForCausalLM.from_pretrained(model, trust_remote_code=True, token = access_token, device_map="auto", offload_folder="offload/")
|
31 |
+
model = PeftModel.from_pretrained(model, peft_model, device_map="auto", offload_folder="offload/")
|
32 |
|
33 |
+
device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')
|
34 |
+
#device = 'cuda:0'
|
35 |
model = model.to(device)
|
36 |
model = model.eval()
|
37 |
|
|
|
401 |
print(model.device)
|
402 |
|
403 |
inputs = tokenizer(pt, return_tensors='pt')
|
404 |
+
inputs = {key: value.to(model.device) for key, value in inputs.items()}
|
405 |
#inputs = {key: value.to(model.device) for key, value in inputs.items()}
|
406 |
+
#inputs = {key: value.to('cuda:0') for key, value in inputs.items()}
|
|
|
407 |
print("Inputs loaded onto devices.")
|
408 |
|
409 |
res = model.generate(
|