Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,8 @@ GGUF версия: https://huggingface.co/pirbis/Vikhr-7B-instruct_0.2-GGUF
|
|
13 |
from peft import PeftModel, PeftConfig
|
14 |
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
15 |
import torch
|
16 |
-
|
|
|
17 |
MODEL_NAME = "Vikhrmodels/Vikhr-7B-instruct_0.2"
|
18 |
DEFAULT_MESSAGE_TEMPLATE = "<s>{role}\n{content}</s>\n"
|
19 |
DEFAULT_SYSTEM_PROMPT = "Ты — Вихрь, русскоязычный автоматический ассистент. Ты разговариваешь с людьми и помогаешь им."
|
@@ -55,10 +56,10 @@ def generate(model, tokenizer, prompt, generation_config):
|
|
55 |
output_ids = output_ids[len(data["input_ids"][0]):]
|
56 |
output = tokenizer.decode(output_ids, skip_special_tokens=True)
|
57 |
return output.strip()
|
58 |
-
|
59 |
#config = PeftConfig.from_pretrained(MODEL_NAME)
|
60 |
model = AutoModelForCausalLM.from_pretrained(
|
61 |
-
|
62 |
load_in_8bit=True,
|
63 |
torch_dtype=torch.float16,
|
64 |
device_map="auto"
|
@@ -67,7 +68,12 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
67 |
model.eval()
|
68 |
|
69 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=False)
|
|
|
70 |
generation_config = GenerationConfig.from_pretrained(MODEL_NAME)
|
|
|
|
|
|
|
|
|
71 |
print(generation_config)
|
72 |
|
73 |
inputs = ["Как тебя зовут?", "Кто такой Колмогоров?"]
|
|
|
13 |
from peft import PeftModel, PeftConfig
|
14 |
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
15 |
import torch
|
16 |
+
import os
|
17 |
+
os.environ['HF_HOME']='.'
|
18 |
MODEL_NAME = "Vikhrmodels/Vikhr-7B-instruct_0.2"
|
19 |
DEFAULT_MESSAGE_TEMPLATE = "<s>{role}\n{content}</s>\n"
|
20 |
DEFAULT_SYSTEM_PROMPT = "Ты — Вихрь, русскоязычный автоматический ассистент. Ты разговариваешь с людьми и помогаешь им."
|
|
|
56 |
output_ids = output_ids[len(data["input_ids"][0]):]
|
57 |
output = tokenizer.decode(output_ids, skip_special_tokens=True)
|
58 |
return output.strip()
|
59 |
+
|
60 |
#config = PeftConfig.from_pretrained(MODEL_NAME)
|
61 |
model = AutoModelForCausalLM.from_pretrained(
|
62 |
+
MODEL_NAME,
|
63 |
load_in_8bit=True,
|
64 |
torch_dtype=torch.float16,
|
65 |
device_map="auto"
|
|
|
68 |
model.eval()
|
69 |
|
70 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=False)
|
71 |
+
|
72 |
generation_config = GenerationConfig.from_pretrained(MODEL_NAME)
|
73 |
+
generation_config.max_length=256
|
74 |
+
generation_config.top_p=0.9
|
75 |
+
generation_config.top_k=30
|
76 |
+
generation_config.do_sample = True
|
77 |
print(generation_config)
|
78 |
|
79 |
inputs = ["Как тебя зовут?", "Кто такой Колмогоров?"]
|