Spaces:
Sleeping
Sleeping
import streamlit as st | |
from llama_cpp import Llama | |
# Title for your app | |
st.title("Llama-3-8B-Physics Master - Chatbot") | |
# Load the model from Hugging Face using llama_cpp | |
def load_model(): | |
# Load the model from the Hugging Face Hub | |
model = Llama.from_pretrained( | |
repo_id="gallen881/Llama-3-8B-Physics_Master-GGUF", | |
filename="unsloth.F16.gguf" # or unsloth.Q4_K_M.gguf for a smaller file | |
) | |
return model | |
# Load the model once and store it in cache | |
model = load_model() | |
# Text input for the user | |
user_input = st.text_area("Enter your message here:") | |
if st.button("Generate Response"): | |
if user_input: | |
# Create chat completion with the model | |
response = model.create_chat_completion( | |
messages=[ | |
{ | |
"role": "user", | |
"content": user_input | |
} | |
] | |
) | |
# Extract the content from the model's response | |
st.write("Model Response:", response['choices'][0]['message']['content']) | |
else: | |
st.write("Please enter a message.") | |