RinInori commited on
Commit
3d6b24a
·
1 Parent(s): d563d4f

Delete InteractiveManual.ipynb

Browse files
Files changed (1) hide show
  1. InteractiveManual.ipynb +0 -356
InteractiveManual.ipynb DELETED
@@ -1,356 +0,0 @@
1
- {
2
- "cells": [
3
- {
4
- "cell_type": "markdown",
5
- "metadata": {
6
- "id": "B3S1AZp6Kc0R"
7
- },
8
- "source": [
9
- "# Installing\n",
10
- "\n",
11
- "If running on Google Colab you will need a Colab Pro+ subscription.\n",
12
- "\n",
13
- "Change the runtime type to a high memory or connect your local machine to colab, launch a vm perhaps.\n",
14
- "\n",
15
- "First, you will need to obtain the LLaMA weights. \n",
16
- "\n",
17
- "You can sign up for the official weights here: https://huggingface.co/docs/transformers/main/model_doc/llama\n",
18
- "\n",
19
- "There are alternative models available on huggingface however. This guide will assume you do not have access to the official weights.\n",
20
- "\n",
21
- "If you do have access to the official weights skip to: Clone the delta weights"
22
- ]
23
- },
24
- {
25
- "cell_type": "markdown",
26
- "metadata": {
27
- "id": "8_MM5GNkKc0T"
28
- },
29
- "source": [
30
- "## Clone the LLaMA weights"
31
- ]
32
- },
33
- {
34
- "cell_type": "code",
35
- "execution_count": null,
36
- "metadata": {
37
- "id": "5TioBc7WKc0U"
38
- },
39
- "outputs": [],
40
- "source": [
41
- "# Setup git lfs\n",
42
- "!git lfs install --skip-smudge --force\n",
43
- "!git lfs env\n",
44
- "!git config filter.lfs.process \"git-lfs filter-process --skip\"\n",
45
- "!git config filter.lfs.smudge \"git-lfs smudge --skip -- %f\""
46
- ]
47
- },
48
- {
49
- "cell_type": "code",
50
- "execution_count": null,
51
- "metadata": {
52
- "id": "B5Q-x5yBKc0V"
53
- },
54
- "outputs": [],
55
- "source": [
56
- "# Cloning the 7b parameter model repo\n",
57
- "!git lfs clone https://huggingface.co/decapoda-research/llama-7b-hf"
58
- ]
59
- },
60
- {
61
- "cell_type": "code",
62
- "execution_count": null,
63
- "metadata": {
64
- "id": "d0r8HBUyKc0V"
65
- },
66
- "outputs": [],
67
- "source": [
68
- "# Cloning the 13b parameter model repo\n",
69
- "!git lfs clone https://huggingface.co/decapoda-research/llama-13b-hf"
70
- ]
71
- },
72
- {
73
- "cell_type": "markdown",
74
- "metadata": {
75
- "id": "QeDIT-a4Kc0V"
76
- },
77
- "source": [
78
- "## Applying the Vicuna delta weights"
79
- ]
80
- },
81
- {
82
- "cell_type": "markdown",
83
- "metadata": {
84
- "id": "p7BGGUICKc0W"
85
- },
86
- "source": [
87
- "### Install PyTorch with CUDA support\n",
88
- "\n",
89
- "If you already have this installed in your environment, you can skip this step."
90
- ]
91
- },
92
- {
93
- "cell_type": "code",
94
- "execution_count": null,
95
- "metadata": {
96
- "id": "IfGE0ALrKc0W"
97
- },
98
- "outputs": [],
99
- "source": [
100
- "# First we need to upgrade setuptools, pip and wheel\n",
101
- "!pip install --upgrade setuptools pip wheel"
102
- ]
103
- },
104
- {
105
- "cell_type": "code",
106
- "execution_count": null,
107
- "metadata": {
108
- "id": "RpvAYAD8Kc0W"
109
- },
110
- "outputs": [],
111
- "source": [
112
- "# For CUDA 11.X:\n",
113
- "!pip install nvidia-cuda-runtime-cu11 --index-url https://pypi.ngc.nvidia.com"
114
- ]
115
- },
116
- {
117
- "cell_type": "code",
118
- "execution_count": null,
119
- "metadata": {
120
- "id": "1oEi1ggaKc0X"
121
- },
122
- "outputs": [],
123
- "source": [
124
- "# For CUDA 12.x\n",
125
- "!pip install nvidia-cuda-runtime-cu12 --index-url https://pypi.ngc.nvidia.com"
126
- ]
127
- },
128
- {
129
- "cell_type": "code",
130
- "execution_count": null,
131
- "metadata": {
132
- "id": "fQYYrn4tKc0X"
133
- },
134
- "outputs": [],
135
- "source": [
136
- "# For PyTorch cu117\n",
137
- "!pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu117"
138
- ]
139
- },
140
- {
141
- "cell_type": "code",
142
- "execution_count": null,
143
- "metadata": {
144
- "id": "G7lnb0sYKc0X"
145
- },
146
- "outputs": [],
147
- "source": [
148
- "# For PyTorch cu118\n",
149
- "!pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118"
150
- ]
151
- },
152
- {
153
- "cell_type": "markdown",
154
- "metadata": {
155
- "id": "oBVyjHuWKc0Y"
156
- },
157
- "source": [
158
- "### Running the Fast-Chat apply delta script"
159
- ]
160
- },
161
- {
162
- "cell_type": "code",
163
- "execution_count": null,
164
- "metadata": {
165
- "id": "a4b11Z8lKc0Y"
166
- },
167
- "outputs": [],
168
- "source": [
169
- "# Install FastChat\n",
170
- "!pip install fschat\n",
171
- "\n",
172
- "# Install the latest main branch of huggingface/transformers\n",
173
- "!pip install git+https://github.com/huggingface/transformers"
174
- ]
175
- },
176
- {
177
- "cell_type": "code",
178
- "execution_count": null,
179
- "metadata": {
180
- "id": "3vJTP1m0Kc0Y"
181
- },
182
- "outputs": [],
183
- "source": [
184
- "import json\n",
185
- "\n",
186
- "def convert_llama_names(model_name: str) -> None:\n",
187
- " \"\"\"Convert LlamaForCausalLM to LlamaForCausalLM and LLaMATokenizer to LlamaTokenizer\"\"\"\n",
188
- " with open(f\"{model_name}/config.json\", \"r\", encoding='utf-8') as f:\n",
189
- " data = f.read()\n",
190
- "\n",
191
- " config = json.loads(data)\n",
192
- " config[\"architectures\"] = [\"LlamaForCausalLM\"]\n",
193
- " with open(f\"{model_name}/config.json\", \"w\", encoding='utf-8') as f:\n",
194
- " json.dump(config, f)\n",
195
- "\n",
196
- "\n",
197
- " with open(f\"{model_name}/tokenizer_config.json\", \"r\", encoding='utf-8') as f:\n",
198
- " data = f.read()\n",
199
- "\n",
200
- " config = json.loads(data)\n",
201
- " config[\"tokenizer_class\"] = \"LlamaTokenizer\"\n",
202
- "\n",
203
- " with open(f\"{model_name}/tokenizer_config.json\", \"w\", encoding='utf-8') as f:\n",
204
- " json.dump(config, f)"
205
- ]
206
- },
207
- {
208
- "cell_type": "code",
209
- "execution_count": null,
210
- "metadata": {
211
- "id": "qyboxxkXKc0Y"
212
- },
213
- "outputs": [],
214
- "source": [
215
- "!git lfs clone https://huggingface.co/lmsys/vicuna-7b-delta-v1.1"
216
- ]
217
- },
218
- {
219
- "cell_type": "code",
220
- "execution_count": null,
221
- "metadata": {
222
- "id": "9tlvfvapKc0Z"
223
- },
224
- "outputs": [],
225
- "source": [
226
- "# 7b Model\n",
227
- "convert_llama_names(\"llama-7b-hf\")\n",
228
- "!python -m fastchat.model.apply_delta --base llama-7b-hf --target vicuna-7b --delta ./vicuna-7b-delta-v1.1"
229
- ]
230
- },
231
- {
232
- "cell_type": "code",
233
- "execution_count": null,
234
- "metadata": {
235
- "id": "5VzAdoySKc0Z"
236
- },
237
- "outputs": [],
238
- "source": [
239
- "!git lfs clone https://huggingface.co/lmsys/vicuna-13b-delta-v1.1"
240
- ]
241
- },
242
- {
243
- "cell_type": "code",
244
- "execution_count": null,
245
- "metadata": {
246
- "id": "jRBc7q5RKc0Z"
247
- },
248
- "outputs": [],
249
- "source": [
250
- "# 13b\n",
251
- "convert_llama_names(\"llama-13b-hf\")\n",
252
- "!python -m fastchat.model.apply_delta --base llama-13b-hf --target vicuna-13b --delta ./vicuna-13b-delta-v1.1"
253
- ]
254
- },
255
- {
256
- "cell_type": "markdown",
257
- "metadata": {
258
- "id": "7n8Ug87RKc0Z"
259
- },
260
- "source": [
261
- "# Installing Auto-Vicuna\n",
262
- "\n",
263
- "Note that running this does not work in colab or the notebook, it is for demonstration purposes only."
264
- ]
265
- },
266
- {
267
- "cell_type": "code",
268
- "execution_count": null,
269
- "metadata": {
270
- "id": "gIlfwNlEKc0Z"
271
- },
272
- "outputs": [],
273
- "source": [
274
- "!pip install auto-vicuna"
275
- ]
276
- },
277
- {
278
- "cell_type": "markdown",
279
- "metadata": {
280
- "id": "0Z5NlQOXKc0Z"
281
- },
282
- "source": [
283
- "# Running Auto-Vicuna"
284
- ]
285
- },
286
- {
287
- "cell_type": "code",
288
- "execution_count": null,
289
- "metadata": {
290
- "id": "sBJiQP7MKc0a"
291
- },
292
- "outputs": [],
293
- "source": [
294
- "!auto_vicuna --vicuna_weights vicuna-7b"
295
- ]
296
- },
297
- {
298
- "cell_type": "markdown",
299
- "metadata": {
300
- "id": "oqqFeF4zKc0a"
301
- },
302
- "source": [
303
- "You can also create a .env file with \n",
304
- "\n",
305
- "```\n",
306
- "VICUNA_WEIGHTS=vicuna-7b\n",
307
- "```\n",
308
- "\n",
309
- "To avoid passing the weights as an arugment."
310
- ]
311
- },
312
- {
313
- "cell_type": "markdown",
314
- "metadata": {
315
- "id": "jiW98C1KKc0a"
316
- },
317
- "source": []
318
- },
319
- {
320
- "cell_type": "markdown",
321
- "metadata": {
322
- "id": "VE8DhmUrKc0a"
323
- },
324
- "source": [
325
- "## Known Issues\n",
326
- "\n",
327
- "If your model keeps talking about random news articles and suchs the `special_tokens_map.json` and `tokenizer_config.json` need to have to stop tokens populated most likely, you can find them in the repo's root dir."
328
- ]
329
- }
330
- ],
331
- "metadata": {
332
- "kernelspec": {
333
- "display_name": ".venv",
334
- "language": "python",
335
- "name": "python3"
336
- },
337
- "language_info": {
338
- "codemirror_mode": {
339
- "name": "ipython",
340
- "version": 3
341
- },
342
- "file_extension": ".py",
343
- "mimetype": "text/x-python",
344
- "name": "python",
345
- "nbconvert_exporter": "python",
346
- "pygments_lexer": "ipython3",
347
- "version": "3.9.13"
348
- },
349
- "orig_nbformat": 4,
350
- "colab": {
351
- "provenance": []
352
- }
353
- },
354
- "nbformat": 4,
355
- "nbformat_minor": 0
356
- }