sotosbarl commited on
Commit
5a2fd52
·
1 Parent(s): a02b0f1

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +25 -2
app.py CHANGED
@@ -1,3 +1,26 @@
1
- import gradio as gr
 
 
2
 
3
- gr.load("models/TinyLlama/TinyLlama-1.1B-Chat-v1.0").launch()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Install transformers from source - only needed for versions <= v4.34
2
+ # pip install git+https://github.com/huggingface/transformers.git
3
+ # pip install accelerate
4
 
5
+ import torch
6
+ from transformers import pipeline
7
+
8
+ pipe = pipeline("text-generation", model="TinyLlama/TinyLlama-1.1B-Chat-v1.0", torch_dtype=torch.bfloat16, device_map="auto")
9
+
10
+ # We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
11
+ messages = [
12
+ {
13
+ "role": "system",
14
+ "content": "You are a friendly chatbot",
15
+ },
16
+ {"role": "user", "content": "I'm 29 years old and I live in Greece. Tellm me how old am I"},
17
+ ]
18
+ prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
19
+ outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
20
+ print(outputs[0]["generated_text"])
21
+ # <|system|>
22
+ # You are a friendly chatbot who always responds in the style of a pirate.</s>
23
+ # <|user|>
24
+ # How many helicopters can a human eat in one sitting?</s>
25
+ # <|assistant|>
26
+ # ...