Tonic commited on
Commit
2368a49
·
1 Parent(s): e13d973

wrap automodel for zerogpu

Browse files
Files changed (1) hide show
  1. langchainapp.py +7 -2
langchainapp.py CHANGED
@@ -43,8 +43,13 @@ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
43
  ### Utils
44
  hf_token, yi_token = load_env_variables()
45
 
46
- tokenizer = AutoTokenizer.from_pretrained(model_name, token=hf_token, trust_remote_code=True)
47
- nvidiamodel = AutoModel.from_pretrained(model_name, token=hf_token, trust_remote_code=True).to(device)
 
 
 
 
 
48
  # nvidiamodel.set_pooling_include_prompt(include_prompt=False)
49
 
50
  def clear_cuda_cache():
 
43
  ### Utils
44
  hf_token, yi_token = load_env_variables()
45
 
46
+ @spaces.GPU
47
+ def load_model():
48
+ tokenizer = AutoTokenizer.from_pretrained(model_name, token=hf_token, trust_remote_code=True)
49
+ return AutoModel.from_pretrained(model_name, token=hf_token, trust_remote_code=True).to(device)
50
+
51
+ # Load model
52
+ nvidiamodel = load_model()
53
  # nvidiamodel.set_pooling_include_prompt(include_prompt=False)
54
 
55
  def clear_cuda_cache():