Christoph Holthaus
commited on
Commit
·
def624a
1
Parent(s):
9307da2
model hf path
Browse files
app.py
CHANGED
@@ -7,13 +7,13 @@ import psutil
|
|
7 |
import os
|
8 |
|
9 |
# load like this - use tne variable everywhere
|
10 |
-
|
11 |
# show warning, when empty and briefs description of how to set it
|
12 |
# also add link to "how to search" with link to bloke by default + example search link + example full value (mistral base?)
|
13 |
# info about ram requirements
|
14 |
|
15 |
# Initing things
|
16 |
-
print(f"debug: init model: {
|
17 |
#llm = Llama(model_path="./model.bin") # LLaMa model
|
18 |
print("! INITING DONE !")
|
19 |
|
@@ -39,7 +39,7 @@ import spaces
|
|
39 |
import torch
|
40 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
41 |
|
42 |
-
DESCRIPTION =
|
43 |
|
44 |
if torch.cuda.is_available():
|
45 |
DESCRIPTION += "\n<p>This space is using CPU only. Use a different one if you want to go fast and use GPU. </p>"
|
|
|
7 |
import os
|
8 |
|
9 |
# load like this - use tne variable everywhere
|
10 |
+
model_hf_path=os.getenv("MODEL_HF_PATH")
|
11 |
# show warning, when empty and briefs description of how to set it
|
12 |
# also add link to "how to search" with link to bloke by default + example search link + example full value (mistral base?)
|
13 |
# info about ram requirements
|
14 |
|
15 |
# Initing things
|
16 |
+
print(f"debug: init model: {model_hf_path}")
|
17 |
#llm = Llama(model_path="./model.bin") # LLaMa model
|
18 |
print("! INITING DONE !")
|
19 |
|
|
|
39 |
import torch
|
40 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
41 |
|
42 |
+
DESCRIPTION = f"# Test model: {model_hf_path}"
|
43 |
|
44 |
if torch.cuda.is_available():
|
45 |
DESCRIPTION += "\n<p>This space is using CPU only. Use a different one if you want to go fast and use GPU. </p>"
|