rodrisouza commited on
Commit
26a4e21
·
verified ·
1 Parent(s): 482e27f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -55,7 +55,7 @@ def load_model(model_name):
55
  tokenizer.pad_token = tokenizer.eos_token
56
  tokenizer.add_special_tokens({'pad_token': tokenizer.eos_token})
57
 
58
- model = AutoModelForCausalLM.from_pretrained(models[model_name], token=hugging_face_token, trust_remote_code=True).to("cuda")
59
  selected_model = model_name
60
  except Exception as e:
61
  print(f"Error loading model {model_name}: {e}")
@@ -89,7 +89,7 @@ def interact(user_input, history, interaction_count):
89
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
90
 
91
  # Generate response using selected model
92
- input_ids = tokenizer(prompt, return_tensors='pt').input_ids.to("cuda")
93
  chat_history_ids = model.generate(input_ids, max_new_tokens=100, pad_token_id=tokenizer.eos_token_id, temperature=0.1)
94
  response = tokenizer.decode(chat_history_ids[:, input_ids.shape[-1]:][0], skip_special_tokens=True)
95
 
 
55
  tokenizer.pad_token = tokenizer.eos_token
56
  tokenizer.add_special_tokens({'pad_token': tokenizer.eos_token})
57
 
58
+ model = AutoModelForCausalLM.from_pretrained(models[model_name], token=hugging_face_token, trust_remote_code=True)
59
  selected_model = model_name
60
  except Exception as e:
61
  print(f"Error loading model {model_name}: {e}")
 
89
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
90
 
91
  # Generate response using selected model
92
+ input_ids = tokenizer(prompt, return_tensors='pt').input_ids
93
  chat_history_ids = model.generate(input_ids, max_new_tokens=100, pad_token_id=tokenizer.eos_token_id, temperature=0.1)
94
  response = tokenizer.decode(chat_history_ids[:, input_ids.shape[-1]:][0], skip_special_tokens=True)
95