Spaces:
Running
Running
File size: 1,252 Bytes
1d777c4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 |
import time
try:
from extensions.telegram_bot.source.generators.abstract_generator import AbstractGenerator
except ImportError:
from source.generators.abstract_generator import AbstractGenerator
class Generator(AbstractGenerator):
model_change_allowed = True # if model changing allowed without stopping.
preset_change_allowed = True # if preset_file changing allowed.
def __init__(self, model_path, n_ctx=2048, seed=0, n_gpu_layers=0):
self.model_path = "like"
def generate_answer(
self,
prompt,
generation_params,
eos_token,
stopping_strings,
default_answer,
turn_template="",
**kwargs,
):
if self.model_path == "like":
time.sleep(2)
answer = "π " + prompt.split("\n")[-1] + " π"
elif self.model_path == "dislike":
time.sleep(2)
answer = "π " + prompt.split("\n")[-1] + " π"
else:
answer = prompt.split("\n")[-1]
return answer
def tokens_count(self, text: str):
return len(text)
def get_model_list(self):
return ["like", "dislike"]
def load_model(self, model_file: str):
self.model_path = model_file
|