SimrusDenuvo commited on
Commit
11dd5a2
·
verified ·
1 Parent(s): b5e2b48

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -14
app.py CHANGED
@@ -1,31 +1,27 @@
1
 
2
- from transformers import AutoTokenizer, AutoModelForCausalLM
3
- import torch
4
  import gradio as gr
5
 
6
- model_name = "sberbank-ai/rugpt3medium_based_on_gpt2"
7
  tokenizer = AutoTokenizer.from_pretrained(model_name)
8
- model = AutoModelForCausalLM.from_pretrained(model_name)
9
 
10
  def generate_response(prompt):
11
- instruction = f"Ответь кратко и ясно на вопрос:\n{prompt.strip()}\nОтвет:"
12
- input_ids = tokenizer.encode(instruction, return_tensors="pt")
13
 
14
  output = model.generate(
15
  input_ids,
16
- max_new_tokens=100,
17
  do_sample=True,
18
- top_k=50,
19
- top_p=0.95,
20
- temperature=0.9,
21
- pad_token_id=tokenizer.eos_token_id,
22
  eos_token_id=tokenizer.eos_token_id
23
  )
24
 
25
  response = tokenizer.decode(output[0], skip_special_tokens=True)
26
-
27
- # Удалим часть промпта из начала
28
- return response.replace(instruction, "").strip()
29
 
30
  iface = gr.Interface(
31
  fn=generate_response,
 
1
 
2
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
 
3
  import gradio as gr
4
 
5
+ model_name = "cointegrated/rut5-small"
6
  tokenizer = AutoTokenizer.from_pretrained(model_name)
7
+ model = AutoModelForSeq2SeqLM.from_pretrained(model_name)
8
 
9
  def generate_response(prompt):
10
+ instruction = f"Ответь кратко: {prompt.strip()}"
11
+ input_ids = tokenizer.encode(instruction, return_tensors="pt", max_length=512, truncation=True)
12
 
13
  output = model.generate(
14
  input_ids,
15
+ max_new_tokens=80,
16
  do_sample=True,
17
+ top_p=0.9,
18
+ temperature=0.7,
19
+ repetition_penalty=1.2,
 
20
  eos_token_id=tokenizer.eos_token_id
21
  )
22
 
23
  response = tokenizer.decode(output[0], skip_special_tokens=True)
24
+ return response.strip()
 
 
25
 
26
  iface = gr.Interface(
27
  fn=generate_response,