ThomasBlumet
commited on
Commit
·
1947635
1
Parent(s):
4e9028a
remove inputs_ids in tokenizer.decode
Browse files
app.py
CHANGED
@@ -28,7 +28,7 @@ def generate_text(input_text,max_new_tokens=512,top_k=50,top_p=0.95,temperature=
|
|
28 |
output = model.generate(input_ids, attention_mask=attention_mask, max_new_tokens=max_new_tokens, top_k=top_k, top_p=top_p, temperature=temperature,do_sample=True)
|
29 |
else:
|
30 |
output = model.generate(input_ids, attention_mask=attention_mask, max_new_tokens=max_new_tokens, top_k=top_k, top_p=top_p, temperature=temperature,do_sample=True)
|
31 |
-
return tokenizer.decode(output[0]
|
32 |
|
33 |
|
34 |
time_story = 0
|
|
|
28 |
output = model.generate(input_ids, attention_mask=attention_mask, max_new_tokens=max_new_tokens, top_k=top_k, top_p=top_p, temperature=temperature,do_sample=True)
|
29 |
else:
|
30 |
output = model.generate(input_ids, attention_mask=attention_mask, max_new_tokens=max_new_tokens, top_k=top_k, top_p=top_p, temperature=temperature,do_sample=True)
|
31 |
+
return tokenizer.decode(output[0], skip_special_tokens=True)
|
32 |
|
33 |
|
34 |
time_story = 0
|