Text Generation
Transformers
Safetensors
Serbian
mistral
mergekit
Merge
text-generation-inference
conversational
datatab commited on
Commit
0635e22
·
verified ·
1 Parent(s): da0da5c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +91 -0
README.md CHANGED
@@ -68,3 +68,94 @@ datasets:
68
  <td><strong>tbd</strong></td>
69
  </tr>
70
  </table>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
68
  <td><strong>tbd</strong></td>
69
  </tr>
70
  </table>
71
+
72
+ ## 💻 Usage
73
+ ```terminal
74
+ !pip -q install git+https://github.com/huggingface/transformers # need to install from github
75
+ !pip install -q datasets loralib sentencepiece
76
+ !pip -q install bitsandbytes accelerate
77
+ ```
78
+
79
+ ```python
80
+ from IPython.display import HTML, display
81
+
82
+ def set_css():
83
+ display(HTML('''
84
+ <style>
85
+ pre {
86
+ white-space: pre-wrap;
87
+ }
88
+ </style>
89
+ '''))
90
+ get_ipython().events.register('pre_run_cell', set_css)
91
+
92
+ ```
93
+
94
+ ```python
95
+ import torch
96
+ import transformers
97
+ from transformers import AutoTokenizer, AutoModelForCausalLM
98
+
99
+ model = AutoModelForCausalLM.from_pretrained(
100
+ "datatab/Yugo60-GPT", torch_dtype="auto"
101
+ )
102
+
103
+ tokenizer = AutoTokenizer.from_pretrained(
104
+ "datatab/Yugo60-GPT", torch_dtype="auto"
105
+ )
106
+
107
+
108
+ ```
109
+
110
+ ```python
111
+ from typing import Optional
112
+ from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
113
+
114
+
115
+ def generate(
116
+ user_content: str, system_content: Optional[str] = ""
117
+ ) -> str:
118
+ system_content = "Ispod je uputstvo koje opisuje zadatak, upareno sa unosom koji pruža dodatni kontekst. Napišite odgovor koji na odgovarajući način kompletira zahtev."
119
+
120
+ messages = [
121
+ {
122
+ "role": "system",
123
+ "content": system_content,
124
+ },
125
+ {"role": "user", "content": user_content},
126
+ ]
127
+
128
+ tokenized_chat = tokenizer.apply_chat_template(
129
+ messages, tokenize=True, add_generation_prompt=True, return_tensors="pt"
130
+ ).to("cuda")
131
+
132
+ text_streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
133
+ output = model.generate(
134
+ tokenized_chat,
135
+ streamer=text_streamer,
136
+ max_new_tokens=2048,
137
+ temperature=0.1,
138
+ repetition_penalty=1.11,
139
+ top_p=0.92,
140
+ top_k=1000,
141
+ pad_token_id=tokenizer.pad_token_id,
142
+ eos_token_id=tokenizer.eos_token_id,
143
+ do_sample=True,
144
+ )
145
+
146
+ generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
147
+
148
+
149
+ ```
150
+
151
+ ```python
152
+ generate("Nabroj mi sve planete suncevog sistemai reci mi koja je najveca planeta")
153
+ ```
154
+
155
+ ```python
156
+ generate("Koja je razlika između lame, vikune i alpake?")
157
+ ```
158
+
159
+ ```python
160
+ generate("Napišite kratku e-poruku Semu Altmanu dajući razloge za GPT-4 otvorenog koda")
161
+ ```