Update README.md
Browse files
README.md
CHANGED
@@ -39,7 +39,9 @@ You can find the experiment on W&B at [this address](https://wandb.ai/dryanfurma
|
|
39 |
|
40 |
## 💻 Usage
|
41 |
|
42 |
-
|
|
|
|
|
43 |
|
44 |
```python
|
45 |
!pip install -qU transformers accelerate bitsandbytes
|
@@ -78,7 +80,9 @@ pipeline = transformers.pipeline(
|
|
78 |
)
|
79 |
```
|
80 |
|
81 |
-
|
|
|
|
|
82 |
|
83 |
```python
|
84 |
messages = [{"role": "user", "content": "What is a large language model?"}]
|
@@ -87,4 +91,8 @@ print("***Prompt:\n", prompt)
|
|
87 |
|
88 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
89 |
print("***Generation:\n", outputs[0]["generated_text"])
|
90 |
-
```
|
|
|
|
|
|
|
|
|
|
39 |
|
40 |
## 💻 Usage
|
41 |
|
42 |
+
<details>
|
43 |
+
|
44 |
+
<summary>Setup</summary>
|
45 |
|
46 |
```python
|
47 |
!pip install -qU transformers accelerate bitsandbytes
|
|
|
80 |
)
|
81 |
```
|
82 |
|
83 |
+
</details>
|
84 |
+
|
85 |
+
Run
|
86 |
|
87 |
```python
|
88 |
messages = [{"role": "user", "content": "What is a large language model?"}]
|
|
|
91 |
|
92 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
93 |
print("***Generation:\n", outputs[0]["generated_text"])
|
94 |
+
```
|
95 |
+
|
96 |
+
Output
|
97 |
+
|
98 |
+
coming
|