Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -7,13 +7,7 @@ import os
|
|
7 |
READ_HF=os.environ["read_hf"]
|
8 |
|
9 |
from unsloth import FastLanguageModel
|
10 |
-
|
11 |
-
model_name = "VanguardAI/CoT_multi_llama_LoRA_4bit", # YOUR MODEL YOU USED FOR TRAINING
|
12 |
-
max_seq_length = 2048,
|
13 |
-
dtype = None,
|
14 |
-
load_in_4bit = True,
|
15 |
-
token= READ_HF
|
16 |
-
)
|
17 |
|
18 |
|
19 |
|
@@ -77,6 +71,13 @@ Category List : ["Dairy & Eggs", "Beverages & Snacks", "Cleaning & Hygiene", "Gr
|
|
77 |
|
78 |
@spaces.GPU()
|
79 |
def chunk_it(inventory_list, user_input_text):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
80 |
FastLanguageModel.for_inference(model) # Enable native 2x faster inference
|
81 |
inputs = tokenizer(
|
82 |
[
|
|
|
7 |
READ_HF=os.environ["read_hf"]
|
8 |
|
9 |
from unsloth import FastLanguageModel
|
10 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
|
12 |
|
13 |
|
|
|
71 |
|
72 |
@spaces.GPU()
|
73 |
def chunk_it(inventory_list, user_input_text):
|
74 |
+
model, tokenizer = FastLanguageModel.from_pretrained(
|
75 |
+
model_name = "VanguardAI/CoT_multi_llama_LoRA_4bit", # YOUR MODEL YOU USED FOR TRAINING
|
76 |
+
max_seq_length = 2048,
|
77 |
+
dtype = None,
|
78 |
+
load_in_4bit = True,
|
79 |
+
token= READ_HF
|
80 |
+
)
|
81 |
FastLanguageModel.for_inference(model) # Enable native 2x faster inference
|
82 |
inputs = tokenizer(
|
83 |
[
|