from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1") tokenizer.pad_token = tokenizer.eos_token # Most LLMs don't have a pad token by default model = AutoModelForCausalLM.from_pretrained( "mistralai/Mistral-7B-v0.1", device_map="auto", load_in_4bit=True ) Generated output is too short/long If not specified in the [~generation.GenerationConfig] file, generate returns up to 20 tokens by default.