Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -32,6 +32,12 @@ if 'hf_token' not in st.session_state:
|
|
32 |
st.session_state.hf_token = None
|
33 |
if 'repo_name' not in st.session_state:
|
34 |
st.session_state.repo_name = None
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
|
36 |
# --- Agent Class ---
|
37 |
class AIAgent:
|
@@ -123,7 +129,7 @@ def chat_interface_with_agent(input_text, agent_name):
|
|
123 |
return f"Agent {agent_name} not found."
|
124 |
|
125 |
# Use a more powerful language model (GPT-3 or similar) for better chat experience
|
126 |
-
model_name = "text-davinci-003" #
|
127 |
try:
|
128 |
model = transformers_pipeline("text-generation", model=model_name)
|
129 |
except EnvironmentError as e:
|
@@ -143,7 +149,7 @@ def chat_interface_with_cluster(input_text, cluster_name):
|
|
143 |
return f"Cluster {cluster_name} not found."
|
144 |
|
145 |
# Use a more powerful language model (GPT-3 or similar) for better chat experience
|
146 |
-
model_name = "text-davinci-003" #
|
147 |
try:
|
148 |
model = transformers_pipeline("text-generation", model=model_name)
|
149 |
except EnvironmentError as e:
|
@@ -260,7 +266,7 @@ def translate_code(code, source_language, target_language):
|
|
260 |
|
261 |
def generate_code(idea):
|
262 |
"""Generates code based on a given idea using a Hugging Face model."""
|
263 |
-
model_name = "bigcode/starcoder" #
|
264 |
try:
|
265 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
266 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
@@ -488,7 +494,10 @@ with tabs[1]:
|
|
488 |
st.subheader("Chat with AI Agents")
|
489 |
selected_agent_or_cluster = st.selectbox("Select an AI agent or cluster", st.session_state.available_agents + st.session_state.available_clusters)
|
490 |
agent_chat_input = st.text_area("Enter your message:")
|
|
|
|
|
491 |
if st.button("Send"):
|
|
|
492 |
if selected_agent_or_cluster in st.session_state.available_agents:
|
493 |
st.session_state.current_agent = selected_agent_or_cluster
|
494 |
st.session_state.current_cluster = None
|
@@ -532,7 +541,10 @@ with tabs[2]:
|
|
532 |
st.subheader("Chat with AI Agents")
|
533 |
selected_agent_or_cluster = st.selectbox("Select an AI agent or cluster", st.session_state.available_agents + st.session_state.available_clusters)
|
534 |
agent_chat_input = st.text_area("Enter your message:")
|
|
|
|
|
535 |
if st.button("Send"):
|
|
|
536 |
if selected_agent_or_cluster in st.session_state.available_agents:
|
537 |
st.session_state.current_agent = selected_agent_or_cluster
|
538 |
st.session_state.current_cluster = None
|
@@ -599,7 +611,10 @@ with tabs[2]:
|
|
599 |
|
600 |
st.write("Code Generation:")
|
601 |
code_idea = st.text_input("Enter your code idea:")
|
|
|
|
|
602 |
if st.button("Generate"):
|
|
|
603 |
generated_code = generate_code(code_idea)
|
604 |
st.code(generated_code, language="python")
|
605 |
|
@@ -608,7 +623,6 @@ with tabs[2]:
|
|
608 |
if st.session_state.current_project:
|
609 |
st.write(f"Current project: {st.session_state.current_project}")
|
610 |
if st.button("Build"):
|
611 |
-
# Implement build logic here
|
612 |
build_dir = build_project(st.session_state.current_project)
|
613 |
st.write(f"Project built successfully! Build directory: {build_dir}")
|
614 |
|
@@ -620,11 +634,9 @@ with tabs[2]:
|
|
620 |
if st.button("Deploy to Hugging Face Spaces"):
|
621 |
st.session_state.hf_token = hf_token
|
622 |
st.session_state.repo_name = repo_name
|
623 |
-
# Implement Hugging Face Spaces deployment logic here
|
624 |
deploy_to_huggingface(build_dir, hf_token, repo_name)
|
625 |
elif deployment_target == "Local":
|
626 |
if st.button("Deploy Locally"):
|
627 |
-
# Implement local deployment logic here
|
628 |
deploy_locally(build_dir)
|
629 |
else:
|
630 |
st.warning("Please select a project first.")
|
|
|
32 |
st.session_state.hf_token = None
|
33 |
if 'repo_name' not in st.session_state:
|
34 |
st.session_state.repo_name = None
|
35 |
+
if 'selected_model' not in st.session_state:
|
36 |
+
st.session_state.selected_model = None
|
37 |
+
if 'selected_code_model' not in st.session_state:
|
38 |
+
st.session_state.selected_code_model = None
|
39 |
+
if 'selected_chat_model' not in st.session_state:
|
40 |
+
st.session_state.selected_chat_model = None
|
41 |
|
42 |
# --- Agent Class ---
|
43 |
class AIAgent:
|
|
|
129 |
return f"Agent {agent_name} not found."
|
130 |
|
131 |
# Use a more powerful language model (GPT-3 or similar) for better chat experience
|
132 |
+
model_name = st.session_state.selected_chat_model or "text-davinci-003" # Default to GPT-3 if not selected
|
133 |
try:
|
134 |
model = transformers_pipeline("text-generation", model=model_name)
|
135 |
except EnvironmentError as e:
|
|
|
149 |
return f"Cluster {cluster_name} not found."
|
150 |
|
151 |
# Use a more powerful language model (GPT-3 or similar) for better chat experience
|
152 |
+
model_name = st.session_state.selected_chat_model or "text-davinci-003" # Default to GPT-3 if not selected
|
153 |
try:
|
154 |
model = transformers_pipeline("text-generation", model=model_name)
|
155 |
except EnvironmentError as e:
|
|
|
266 |
|
267 |
def generate_code(idea):
|
268 |
"""Generates code based on a given idea using a Hugging Face model."""
|
269 |
+
model_name = st.session_state.selected_code_model or "bigcode/starcoder" # Default to Starcoder if not selected
|
270 |
try:
|
271 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
272 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
|
494 |
st.subheader("Chat with AI Agents")
|
495 |
selected_agent_or_cluster = st.selectbox("Select an AI agent or cluster", st.session_state.available_agents + st.session_state.available_clusters)
|
496 |
agent_chat_input = st.text_area("Enter your message:")
|
497 |
+
chat_model_options = ["text-davinci-003", "gpt-3.5-turbo"] # Add more chat models as needed
|
498 |
+
selected_chat_model = st.selectbox("Select a chat model", chat_model_options)
|
499 |
if st.button("Send"):
|
500 |
+
st.session_state.selected_chat_model = selected_chat_model
|
501 |
if selected_agent_or_cluster in st.session_state.available_agents:
|
502 |
st.session_state.current_agent = selected_agent_or_cluster
|
503 |
st.session_state.current_cluster = None
|
|
|
541 |
st.subheader("Chat with AI Agents")
|
542 |
selected_agent_or_cluster = st.selectbox("Select an AI agent or cluster", st.session_state.available_agents + st.session_state.available_clusters)
|
543 |
agent_chat_input = st.text_area("Enter your message:")
|
544 |
+
chat_model_options = ["text-davinci-003", "gpt-3.5-turbo"] # Add more chat models as needed
|
545 |
+
selected_chat_model = st.selectbox("Select a chat model", chat_model_options)
|
546 |
if st.button("Send"):
|
547 |
+
st.session_state.selected_chat_model = selected_chat_model
|
548 |
if selected_agent_or_cluster in st.session_state.available_agents:
|
549 |
st.session_state.current_agent = selected_agent_or_cluster
|
550 |
st.session_state.current_cluster = None
|
|
|
611 |
|
612 |
st.write("Code Generation:")
|
613 |
code_idea = st.text_input("Enter your code idea:")
|
614 |
+
code_model_options = ["bigcode/starcoder", "google/flan-t5-xl"] # Add more code models as needed
|
615 |
+
selected_code_model = st.selectbox("Select a code generation model", code_model_options)
|
616 |
if st.button("Generate"):
|
617 |
+
st.session_state.selected_code_model = selected_code_model
|
618 |
generated_code = generate_code(code_idea)
|
619 |
st.code(generated_code, language="python")
|
620 |
|
|
|
623 |
if st.session_state.current_project:
|
624 |
st.write(f"Current project: {st.session_state.current_project}")
|
625 |
if st.button("Build"):
|
|
|
626 |
build_dir = build_project(st.session_state.current_project)
|
627 |
st.write(f"Project built successfully! Build directory: {build_dir}")
|
628 |
|
|
|
634 |
if st.button("Deploy to Hugging Face Spaces"):
|
635 |
st.session_state.hf_token = hf_token
|
636 |
st.session_state.repo_name = repo_name
|
|
|
637 |
deploy_to_huggingface(build_dir, hf_token, repo_name)
|
638 |
elif deployment_target == "Local":
|
639 |
if st.button("Deploy Locally"):
|
|
|
640 |
deploy_locally(build_dir)
|
641 |
else:
|
642 |
st.warning("Please select a project first.")
|