Spaces:
Build error
Build error
wrap automodel for zerogpu
Browse files- langchainapp.py +7 -2
langchainapp.py
CHANGED
@@ -43,8 +43,13 @@ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
43 |
### Utils
|
44 |
hf_token, yi_token = load_env_variables()
|
45 |
|
46 |
-
|
47 |
-
|
|
|
|
|
|
|
|
|
|
|
48 |
# nvidiamodel.set_pooling_include_prompt(include_prompt=False)
|
49 |
|
50 |
def clear_cuda_cache():
|
|
|
43 |
### Utils
|
44 |
hf_token, yi_token = load_env_variables()
|
45 |
|
46 |
+
@spaces.GPU
|
47 |
+
def load_model():
|
48 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name, token=hf_token, trust_remote_code=True)
|
49 |
+
return AutoModel.from_pretrained(model_name, token=hf_token, trust_remote_code=True).to(device)
|
50 |
+
|
51 |
+
# Load model
|
52 |
+
nvidiamodel = load_model()
|
53 |
# nvidiamodel.set_pooling_include_prompt(include_prompt=False)
|
54 |
|
55 |
def clear_cuda_cache():
|