File size: 514 Bytes
5fa1a76 |
1 2 3 4 5 6 7 8 9 10 11 |
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1", padding_side="left") tokenizer.pad_token = tokenizer.eos_token # Most LLMs don't have a pad token by default model_inputs = tokenizer( ["1, 2, 3", "A, B, C, D, E"], padding=True, return_tensors="pt" ).to("cuda") generated_ids = model.generate(**model_inputs) tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0] '1, 2, 3, 4, 5, 6,' Wrong prompt Some models and tasks expect a certain input prompt format to work properly. |