Update app.py
Browse files
app.py
CHANGED
|
@@ -7,13 +7,7 @@ import os
|
|
| 7 |
READ_HF=os.environ["read_hf"]
|
| 8 |
|
| 9 |
from unsloth import FastLanguageModel
|
| 10 |
-
|
| 11 |
-
model_name = "VanguardAI/CoT_multi_llama_LoRA_4bit", # YOUR MODEL YOU USED FOR TRAINING
|
| 12 |
-
max_seq_length = 2048,
|
| 13 |
-
dtype = None,
|
| 14 |
-
load_in_4bit = True,
|
| 15 |
-
token= READ_HF
|
| 16 |
-
)
|
| 17 |
|
| 18 |
|
| 19 |
|
|
@@ -77,6 +71,13 @@ Category List : ["Dairy & Eggs", "Beverages & Snacks", "Cleaning & Hygiene", "Gr
|
|
| 77 |
|
| 78 |
@spaces.GPU()
|
| 79 |
def chunk_it(inventory_list, user_input_text):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 80 |
FastLanguageModel.for_inference(model) # Enable native 2x faster inference
|
| 81 |
inputs = tokenizer(
|
| 82 |
[
|
|
|
|
| 7 |
READ_HF=os.environ["read_hf"]
|
| 8 |
|
| 9 |
from unsloth import FastLanguageModel
|
| 10 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
|
| 13 |
|
|
|
|
| 71 |
|
| 72 |
@spaces.GPU()
|
| 73 |
def chunk_it(inventory_list, user_input_text):
|
| 74 |
+
model, tokenizer = FastLanguageModel.from_pretrained(
|
| 75 |
+
model_name = "VanguardAI/CoT_multi_llama_LoRA_4bit", # YOUR MODEL YOU USED FOR TRAINING
|
| 76 |
+
max_seq_length = 2048,
|
| 77 |
+
dtype = None,
|
| 78 |
+
load_in_4bit = True,
|
| 79 |
+
token= READ_HF
|
| 80 |
+
)
|
| 81 |
FastLanguageModel.for_inference(model) # Enable native 2x faster inference
|
| 82 |
inputs = tokenizer(
|
| 83 |
[
|