metastable-void
commited on
update
Browse files- app.py +10 -3
- pyproject.toml +8 -7
- requirements.txt +21 -7
app.py
CHANGED
|
@@ -8,6 +8,7 @@ import gradio as gr
|
|
| 8 |
import spaces
|
| 9 |
import torch
|
| 10 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer, pipeline
|
|
|
|
| 11 |
|
| 12 |
DESCRIPTION = "# chat-1"
|
| 13 |
|
|
@@ -22,10 +23,16 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "32768"))
|
|
| 22 |
if torch.cuda.is_available():
|
| 23 |
model_id = "vericava/llm-jp-3-1.8b-instruct-lora-vericava7-llama"
|
| 24 |
base_model_id = "llm-jp/llm-jp-3-1.8b-instruct"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
my_pipeline=pipeline(
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
)
|
| 30 |
my_pipeline.tokenizer.chat_template = "{{bos_token}}{% for message in messages %}{% if message['role'] == 'user' %}{{ '\\n\\n### 前の投稿:\\n' + message['content'] + '' }}{% elif message['role'] == 'system' %}{{ '以下は、SNS上の投稿です。あなたはSNSの投稿生成botとして、次に続く投稿を考えなさい。説明はせず、投稿の内容のみを鉤括弧をつけずに答えよ。' }}{% elif message['role'] == 'assistant' %}{{ '\\n\\n### 次の投稿:\\n' + message['content'] + eos_token }}{% endif %}{% if loop.last and add_generation_prompt %}{{ '\\n\\n### 次の投稿:\\n' }}{% endif %}{% endfor %}"
|
| 31 |
|
|
|
|
| 8 |
import spaces
|
| 9 |
import torch
|
| 10 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer, pipeline
|
| 11 |
+
from peft import PeftModel
|
| 12 |
|
| 13 |
DESCRIPTION = "# chat-1"
|
| 14 |
|
|
|
|
| 23 |
if torch.cuda.is_available():
|
| 24 |
model_id = "vericava/llm-jp-3-1.8b-instruct-lora-vericava7-llama"
|
| 25 |
base_model_id = "llm-jp/llm-jp-3-1.8b-instruct"
|
| 26 |
+
tokenizer = AutoTokenizer.from_pretrained(base_model_id, trust_remote_code=True)
|
| 27 |
+
base_model = AutoModelForCausalLM.from_pretrained(
|
| 28 |
+
base_model_id,
|
| 29 |
+
trust_remote_code=True,
|
| 30 |
+
)
|
| 31 |
+
model = PeftModel.from_pretrained(base_model, model_id, trust_remote_code=True)
|
| 32 |
my_pipeline=pipeline(
|
| 33 |
+
task="text-generation",
|
| 34 |
+
model=model,
|
| 35 |
+
tokenizer=tokenizer,
|
| 36 |
)
|
| 37 |
my_pipeline.tokenizer.chat_template = "{{bos_token}}{% for message in messages %}{% if message['role'] == 'user' %}{{ '\\n\\n### 前の投稿:\\n' + message['content'] + '' }}{% elif message['role'] == 'system' %}{{ '以下は、SNS上の投稿です。あなたはSNSの投稿生成botとして、次に続く投稿を考えなさい。説明はせず、投稿の内容のみを鉤括弧をつけずに答えよ。' }}{% elif message['role'] == 'assistant' %}{{ '\\n\\n### 次の投稿:\\n' + message['content'] + eos_token }}{% endif %}{% if loop.last and add_generation_prompt %}{{ '\\n\\n### 次の投稿:\\n' }}{% endif %}{% endfor %}"
|
| 38 |
|
pyproject.toml
CHANGED
|
@@ -1,12 +1,13 @@
|
|
| 1 |
[project]
|
| 2 |
dependencies = [
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
|
|
|
| 10 |
]
|
| 11 |
description = ""
|
| 12 |
name = "chat-1"
|
|
|
|
| 1 |
[project]
|
| 2 |
dependencies = [
|
| 3 |
+
"accelerate>=1.2.1",
|
| 4 |
+
"bitsandbytes>=0.45.0",
|
| 5 |
+
"gradio>=5.9.1",
|
| 6 |
+
"hf-transfer>=0.1.8",
|
| 7 |
+
"spaces>=0.31.1",
|
| 8 |
+
"torch==2.4.0",
|
| 9 |
+
"transformers>=4.48.3",
|
| 10 |
+
"peft==0.14.0",
|
| 11 |
]
|
| 12 |
description = ""
|
| 13 |
name = "chat-1"
|
requirements.txt
CHANGED
|
@@ -1,7 +1,9 @@
|
|
| 1 |
# This file was autogenerated by uv via the following command:
|
| 2 |
# uv pip compile pyproject.toml -o requirements.txt
|
| 3 |
accelerate==1.2.1
|
| 4 |
-
# via
|
|
|
|
|
|
|
| 5 |
aiofiles==23.2.1
|
| 6 |
# via gradio
|
| 7 |
annotated-types==0.7.0
|
|
@@ -12,7 +14,7 @@ anyio==4.7.0
|
|
| 12 |
# httpx
|
| 13 |
# starlette
|
| 14 |
bitsandbytes==0.45.0
|
| 15 |
-
# via
|
| 16 |
certifi==2024.12.14
|
| 17 |
# via
|
| 18 |
# httpcore
|
|
@@ -41,7 +43,7 @@ fsspec==2024.12.0
|
|
| 41 |
# torch
|
| 42 |
gradio==5.9.1
|
| 43 |
# via
|
| 44 |
-
#
|
| 45 |
# spaces
|
| 46 |
gradio-client==1.5.2
|
| 47 |
# via gradio
|
|
@@ -50,7 +52,7 @@ h11==0.14.0
|
|
| 50 |
# httpcore
|
| 51 |
# uvicorn
|
| 52 |
hf-transfer==0.1.8
|
| 53 |
-
# via
|
| 54 |
httpcore==1.0.7
|
| 55 |
# via httpx
|
| 56 |
httpx==0.28.1
|
|
@@ -64,6 +66,7 @@ huggingface-hub==0.27.0
|
|
| 64 |
# accelerate
|
| 65 |
# gradio
|
| 66 |
# gradio-client
|
|
|
|
| 67 |
# tokenizers
|
| 68 |
# transformers
|
| 69 |
idna==3.10
|
|
@@ -93,6 +96,7 @@ numpy==2.2.1
|
|
| 93 |
# bitsandbytes
|
| 94 |
# gradio
|
| 95 |
# pandas
|
|
|
|
| 96 |
# transformers
|
| 97 |
nvidia-cublas-cu12==12.1.3.1
|
| 98 |
# via
|
|
@@ -133,15 +137,19 @@ packaging==24.2
|
|
| 133 |
# gradio
|
| 134 |
# gradio-client
|
| 135 |
# huggingface-hub
|
|
|
|
| 136 |
# spaces
|
| 137 |
# transformers
|
| 138 |
pandas==2.2.3
|
| 139 |
# via gradio
|
|
|
|
|
|
|
| 140 |
pillow==11.1.0
|
| 141 |
# via gradio
|
| 142 |
psutil==5.9.8
|
| 143 |
# via
|
| 144 |
# accelerate
|
|
|
|
| 145 |
# spaces
|
| 146 |
pydantic==2.10.4
|
| 147 |
# via
|
|
@@ -165,6 +173,7 @@ pyyaml==6.0.2
|
|
| 165 |
# accelerate
|
| 166 |
# gradio
|
| 167 |
# huggingface-hub
|
|
|
|
| 168 |
# transformers
|
| 169 |
regex==2024.11.6
|
| 170 |
# via transformers
|
|
@@ -182,6 +191,7 @@ safehttpx==0.1.6
|
|
| 182 |
safetensors==0.4.5
|
| 183 |
# via
|
| 184 |
# accelerate
|
|
|
|
| 185 |
# transformers
|
| 186 |
semantic-version==2.10.0
|
| 187 |
# via gradio
|
|
@@ -192,7 +202,7 @@ six==1.17.0
|
|
| 192 |
sniffio==1.3.1
|
| 193 |
# via anyio
|
| 194 |
spaces==0.31.1
|
| 195 |
-
# via
|
| 196 |
starlette==0.41.3
|
| 197 |
# via
|
| 198 |
# fastapi
|
|
@@ -205,15 +215,19 @@ tomlkit==0.13.2
|
|
| 205 |
# via gradio
|
| 206 |
torch==2.4.0
|
| 207 |
# via
|
| 208 |
-
#
|
| 209 |
# accelerate
|
| 210 |
# bitsandbytes
|
|
|
|
| 211 |
tqdm==4.67.1
|
| 212 |
# via
|
| 213 |
# huggingface-hub
|
|
|
|
| 214 |
# transformers
|
| 215 |
transformers==4.50.1
|
| 216 |
-
# via
|
|
|
|
|
|
|
| 217 |
triton==3.0.0
|
| 218 |
# via torch
|
| 219 |
typer==0.15.1
|
|
|
|
| 1 |
# This file was autogenerated by uv via the following command:
|
| 2 |
# uv pip compile pyproject.toml -o requirements.txt
|
| 3 |
accelerate==1.2.1
|
| 4 |
+
# via
|
| 5 |
+
# chat-1 (pyproject.toml)
|
| 6 |
+
# peft
|
| 7 |
aiofiles==23.2.1
|
| 8 |
# via gradio
|
| 9 |
annotated-types==0.7.0
|
|
|
|
| 14 |
# httpx
|
| 15 |
# starlette
|
| 16 |
bitsandbytes==0.45.0
|
| 17 |
+
# via chat-1 (pyproject.toml)
|
| 18 |
certifi==2024.12.14
|
| 19 |
# via
|
| 20 |
# httpcore
|
|
|
|
| 43 |
# torch
|
| 44 |
gradio==5.9.1
|
| 45 |
# via
|
| 46 |
+
# chat-1 (pyproject.toml)
|
| 47 |
# spaces
|
| 48 |
gradio-client==1.5.2
|
| 49 |
# via gradio
|
|
|
|
| 52 |
# httpcore
|
| 53 |
# uvicorn
|
| 54 |
hf-transfer==0.1.8
|
| 55 |
+
# via chat-1 (pyproject.toml)
|
| 56 |
httpcore==1.0.7
|
| 57 |
# via httpx
|
| 58 |
httpx==0.28.1
|
|
|
|
| 66 |
# accelerate
|
| 67 |
# gradio
|
| 68 |
# gradio-client
|
| 69 |
+
# peft
|
| 70 |
# tokenizers
|
| 71 |
# transformers
|
| 72 |
idna==3.10
|
|
|
|
| 96 |
# bitsandbytes
|
| 97 |
# gradio
|
| 98 |
# pandas
|
| 99 |
+
# peft
|
| 100 |
# transformers
|
| 101 |
nvidia-cublas-cu12==12.1.3.1
|
| 102 |
# via
|
|
|
|
| 137 |
# gradio
|
| 138 |
# gradio-client
|
| 139 |
# huggingface-hub
|
| 140 |
+
# peft
|
| 141 |
# spaces
|
| 142 |
# transformers
|
| 143 |
pandas==2.2.3
|
| 144 |
# via gradio
|
| 145 |
+
peft==0.14.0
|
| 146 |
+
# via chat-1 (pyproject.toml)
|
| 147 |
pillow==11.1.0
|
| 148 |
# via gradio
|
| 149 |
psutil==5.9.8
|
| 150 |
# via
|
| 151 |
# accelerate
|
| 152 |
+
# peft
|
| 153 |
# spaces
|
| 154 |
pydantic==2.10.4
|
| 155 |
# via
|
|
|
|
| 173 |
# accelerate
|
| 174 |
# gradio
|
| 175 |
# huggingface-hub
|
| 176 |
+
# peft
|
| 177 |
# transformers
|
| 178 |
regex==2024.11.6
|
| 179 |
# via transformers
|
|
|
|
| 191 |
safetensors==0.4.5
|
| 192 |
# via
|
| 193 |
# accelerate
|
| 194 |
+
# peft
|
| 195 |
# transformers
|
| 196 |
semantic-version==2.10.0
|
| 197 |
# via gradio
|
|
|
|
| 202 |
sniffio==1.3.1
|
| 203 |
# via anyio
|
| 204 |
spaces==0.31.1
|
| 205 |
+
# via chat-1 (pyproject.toml)
|
| 206 |
starlette==0.41.3
|
| 207 |
# via
|
| 208 |
# fastapi
|
|
|
|
| 215 |
# via gradio
|
| 216 |
torch==2.4.0
|
| 217 |
# via
|
| 218 |
+
# chat-1 (pyproject.toml)
|
| 219 |
# accelerate
|
| 220 |
# bitsandbytes
|
| 221 |
+
# peft
|
| 222 |
tqdm==4.67.1
|
| 223 |
# via
|
| 224 |
# huggingface-hub
|
| 225 |
+
# peft
|
| 226 |
# transformers
|
| 227 |
transformers==4.50.1
|
| 228 |
+
# via
|
| 229 |
+
# chat-1 (pyproject.toml)
|
| 230 |
+
# peft
|
| 231 |
triton==3.0.0
|
| 232 |
# via torch
|
| 233 |
typer==0.15.1
|