zhengr commited on
Commit
e97c212
·
verified ·
1 Parent(s): 8aa9500

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -21,17 +21,17 @@ access_token = os.environ["TOKEN"]
21
  model = "NousResearch/Llama-2-7b-hf"
22
  peft_model = "FinGPT/fingpt-forecaster_sz50_llama2-7B_lora"
23
 
24
- tokenizer = AutoTokenizer.from_pretrained(model, token = access_token, device_map="cuda:0", trust_remote_code=True)
25
  tokenizer.pad_token = tokenizer.eos_token
26
  tokenizer.padding_side = "right"
27
 
28
  streamer = TextStreamer(tokenizer)
29
 
30
- model = AutoModelForCausalLM.from_pretrained(model, trust_remote_code=True, token = access_token, device_map="cuda:0", load_in_8bit=True, offload_folder="offload/")
31
- model = PeftModel.from_pretrained(model, peft_model, device_map="cuda:0", offload_folder="offload/")
32
 
33
- #device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')
34
- device = 'cuda:0'
35
  model = model.to(device)
36
  model = model.eval()
37
 
@@ -401,9 +401,9 @@ def ask(symbol, weeks_before, withbasic):
401
  print(model.device)
402
 
403
  inputs = tokenizer(pt, return_tensors='pt')
 
404
  #inputs = {key: value.to(model.device) for key, value in inputs.items()}
405
- #inputs = {key: value.to(model.device) for key, value in inputs.items()}
406
- inputs = {key: value.to('cuda:0') for key, value in inputs.items()}
407
  print("Inputs loaded onto devices.")
408
 
409
  res = model.generate(
 
21
  model = "NousResearch/Llama-2-7b-hf"
22
  peft_model = "FinGPT/fingpt-forecaster_sz50_llama2-7B_lora"
23
 
24
+ tokenizer = AutoTokenizer.from_pretrained(model, token = access_token, device_map="auto", trust_remote_code=True)
25
  tokenizer.pad_token = tokenizer.eos_token
26
  tokenizer.padding_side = "right"
27
 
28
  streamer = TextStreamer(tokenizer)
29
 
30
+ model = AutoModelForCausalLM.from_pretrained(model, trust_remote_code=True, token = access_token, device_map="auto", offload_folder="offload/")
31
+ model = PeftModel.from_pretrained(model, peft_model, device_map="auto", offload_folder="offload/")
32
 
33
+ device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')
34
+ #device = 'cuda:0'
35
  model = model.to(device)
36
  model = model.eval()
37
 
 
401
  print(model.device)
402
 
403
  inputs = tokenizer(pt, return_tensors='pt')
404
+ inputs = {key: value.to(model.device) for key, value in inputs.items()}
405
  #inputs = {key: value.to(model.device) for key, value in inputs.items()}
406
+ #inputs = {key: value.to('cuda:0') for key, value in inputs.items()}
 
407
  print("Inputs loaded onto devices.")
408
 
409
  res = model.generate(