Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -17,7 +17,7 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
|
17 |
|
18 |
#Transformers Code
|
19 |
if torch.cuda.is_available():
|
20 |
-
model_id = "BenBranyon/zephyr-sumbot-all-songs"
|
21 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
|
22 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
23 |
tokenizer.use_default_system_prompt = False
|
@@ -71,7 +71,7 @@ def generate(
|
|
71 |
conversation.append({"role": "system", "content": system_prompt})
|
72 |
for user, assistant in chat_history:
|
73 |
conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
|
74 |
-
conversation.append({"role": "user", "content": "
|
75 |
|
76 |
input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt")
|
77 |
if input_ids.shape[1] > MAX_INPUT_TOKEN_LENGTH:
|
|
|
17 |
|
18 |
#Transformers Code
|
19 |
if torch.cuda.is_available():
|
20 |
+
model_id = "BenBranyon/zephyr-sumbot-all-songs-split"
|
21 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
|
22 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
23 |
tokenizer.use_default_system_prompt = False
|
|
|
71 |
conversation.append({"role": "system", "content": system_prompt})
|
72 |
for user, assistant in chat_history:
|
73 |
conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
|
74 |
+
conversation.append({"role": "user", "content": "Generate rap lyrics in the style of the artist Sumkilla about " + message})
|
75 |
|
76 |
input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt")
|
77 |
if input_ids.shape[1] > MAX_INPUT_TOKEN_LENGTH:
|