Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -13,11 +13,11 @@ DEFAULT_MAX_NEW_TOKENS = 512
|
|
13 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
14 |
|
15 |
#Inference API Code
|
16 |
-
client = InferenceClient("BenBranyon/zephyr-sumbot-all-songs-large")
|
17 |
|
18 |
#Transformers Code
|
19 |
if torch.cuda.is_available():
|
20 |
-
model_id = "BenBranyon/zephyr-sumbot-all-songs"
|
21 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
|
22 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
23 |
tokenizer.use_default_system_prompt = False
|
|
|
13 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
14 |
|
15 |
#Inference API Code
|
16 |
+
#client = InferenceClient("BenBranyon/zephyr-sumbot-all-songs-large")
|
17 |
|
18 |
#Transformers Code
|
19 |
if torch.cuda.is_available():
|
20 |
+
model_id = "BenBranyon/zephyr-sumbot-all-songs-large"
|
21 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
|
22 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
23 |
tokenizer.use_default_system_prompt = False
|