Asman2010 commited on
Commit
3256302
·
verified ·
1 Parent(s): 92ccad4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -15
app.py CHANGED
@@ -16,12 +16,6 @@ if not GROQ_API_KEY:
16
  # Connect to Groq
17
  client = Groq(api_key=GROQ_API_KEY)
18
 
19
- #models' endpoints:
20
- GROQ_MODEL_ENDPOINT70B = os.environ.get(GROQ_API_KEY, model='llama3-70b-8192')
21
- GROQ_MODEL_ENDPOINT8B = os.environ.get(GROQ_API_KEY, model='llama3-8b-8192')
22
- PRE_PROMPT = "You are a helpful assistant. You do not respond as 'User' or pretend to be 'User'. You only respond once as Assistant."
23
- #Auth0 for auth
24
-
25
  # Set up Streamlit app
26
  st.set_page_config(page_title="LLaMA 3x", page_icon="🦙", layout="wide")
27
 
@@ -64,14 +58,7 @@ def render_app():
64
  st.session_state['max_seq_len'] = 512
65
  if 'pre_prompt' not in st.session_state:
66
  st.session_state['pre_prompt'] = PRE_PROMPT
67
-
68
-
69
- #Dropdown menu to select the model edpoint:
70
- selected_option = st.sidebar.selectbox('Choose a LLaMA2 model:', ['LLaMA3 70B', 'LLaMA3 8B'], key='model')
71
- if selected_option == 'LLaMA-3 70B':
72
- st.session_state['llm'] = GROQ_MODEL_ENDPOINT70B
73
- else:
74
- st.session_state['llm'] = GROQ_MODEL_ENDPOINT8B
75
 
76
  # Model hyperparameters
77
  st.session_state['temperature'] = st.sidebar.slider('Temperature:', min_value=0.01, max_value=5.0, value=0.1, step=0.01)
@@ -129,7 +116,7 @@ def render_app():
129
  messages = [{"role": msg["role"], "content": msg["content"]} for msg in st.session_state.chat_dialogue]
130
  chat_completion = client.chat.completions.create(
131
  messages=messages,
132
- model=selected_option,
133
  temperature=st.session_state['temperature'],
134
  top_p=st.session_state['top_p'],
135
  max_tokens=st.session_state['max_seq_len']
 
16
  # Connect to Groq
17
  client = Groq(api_key=GROQ_API_KEY)
18
 
 
 
 
 
 
 
19
  # Set up Streamlit app
20
  st.set_page_config(page_title="LLaMA 3x", page_icon="🦙", layout="wide")
21
 
 
58
  st.session_state['max_seq_len'] = 512
59
  if 'pre_prompt' not in st.session_state:
60
  st.session_state['pre_prompt'] = PRE_PROMPT
61
+
 
 
 
 
 
 
 
62
 
63
  # Model hyperparameters
64
  st.session_state['temperature'] = st.sidebar.slider('Temperature:', min_value=0.01, max_value=5.0, value=0.1, step=0.01)
 
116
  messages = [{"role": msg["role"], "content": msg["content"]} for msg in st.session_state.chat_dialogue]
117
  chat_completion = client.chat.completions.create(
118
  messages=messages,
119
+ model="llama3-70b-8192",
120
  temperature=st.session_state['temperature'],
121
  top_p=st.session_state['top_p'],
122
  max_tokens=st.session_state['max_seq_len']