Spaces:
Configuration error
Configuration error
Update app.py
Browse files
app.py
CHANGED
@@ -55,7 +55,7 @@ def load_model(model_name):
|
|
55 |
tokenizer.pad_token = tokenizer.eos_token
|
56 |
tokenizer.add_special_tokens({'pad_token': tokenizer.eos_token})
|
57 |
|
58 |
-
model = AutoModelForCausalLM.from_pretrained(models[model_name], token=hugging_face_token, trust_remote_code=True)
|
59 |
selected_model = model_name
|
60 |
except Exception as e:
|
61 |
print(f"Error loading model {model_name}: {e}")
|
@@ -89,7 +89,7 @@ def interact(user_input, history, interaction_count):
|
|
89 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
90 |
|
91 |
# Generate response using selected model
|
92 |
-
input_ids = tokenizer(prompt, return_tensors='pt').input_ids
|
93 |
chat_history_ids = model.generate(input_ids, max_new_tokens=100, pad_token_id=tokenizer.eos_token_id, temperature=0.1)
|
94 |
response = tokenizer.decode(chat_history_ids[:, input_ids.shape[-1]:][0], skip_special_tokens=True)
|
95 |
|
|
|
55 |
tokenizer.pad_token = tokenizer.eos_token
|
56 |
tokenizer.add_special_tokens({'pad_token': tokenizer.eos_token})
|
57 |
|
58 |
+
model = AutoModelForCausalLM.from_pretrained(models[model_name], token=hugging_face_token, trust_remote_code=True)
|
59 |
selected_model = model_name
|
60 |
except Exception as e:
|
61 |
print(f"Error loading model {model_name}: {e}")
|
|
|
89 |
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
90 |
|
91 |
# Generate response using selected model
|
92 |
+
input_ids = tokenizer(prompt, return_tensors='pt').input_ids
|
93 |
chat_history_ids = model.generate(input_ids, max_new_tokens=100, pad_token_id=tokenizer.eos_token_id, temperature=0.1)
|
94 |
response = tokenizer.decode(chat_history_ids[:, input_ids.shape[-1]:][0], skip_special_tokens=True)
|
95 |
|