Akjava commited on
Commit
217bb15
·
verified ·
1 Parent(s): 4df7b60

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +27 -3
app.py CHANGED
@@ -106,6 +106,29 @@ hf_hub_download(
106
 
107
  retriever_tool = RetrieverTool(docs_processed)
108
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
109
  # based https://github.com/huggingface/smolagents/pull/450
110
  # almost overwrite with https://huggingface.co/spaces/sitammeur/Gemma-llamacpp
111
  class LlamaCppModel(Model):
@@ -207,13 +230,14 @@ class LlamaCppModel(Model):
207
  agent = LlamaCppAgent(
208
  provider,
209
  system_prompt=f"{system_message}",
210
- predefined_messages_formatter_type=MessagesFormatterType.GEMMA_2,
 
211
  debug_output=True,
212
  )
213
- temperature = 0.7
214
  top_k=40
215
  top_p=0.95
216
- max_tokens=1024
217
  repeat_penalty=1.1
218
  settings = provider.get_provider_default_settings()
219
  settings.temperature = temperature
 
106
 
107
  retriever_tool = RetrieverTool(docs_processed)
108
 
109
+ # Define the prompt markers for Gemma 3
110
+ gemma_3_prompt_markers = {
111
+ Roles.system: PromptMarkers("", "\n"), # System prompt should be included within user message
112
+ Roles.user: PromptMarkers("<start_of_turn>user\n", "<end_of_turn>\n"),
113
+ Roles.assistant: PromptMarkers("<start_of_turn>model\n", "<end_of_turn>\n"),
114
+ Roles.tool: PromptMarkers("", ""), # If you need tool support
115
+ }
116
+
117
+
118
+
119
+
120
+
121
+ # Create the formatter
122
+ gemma_3_formatter = MessagesFormatter(
123
+ pre_prompt="", # No pre-prompt
124
+ prompt_markers=gemma_3_prompt_markers,
125
+ include_sys_prompt_in_first_user_message=True, # Include system prompt in first user message
126
+ default_stop_sequences=["<end_of_turn>", "<start_of_turn>"],
127
+ strip_prompt=False, # Don't strip whitespace from the prompt
128
+ bos_token="<bos>", # Beginning of sequence token for Gemma 3
129
+ eos_token="<eos>", # End of sequence token for Gemma 3
130
+ )
131
+
132
  # based https://github.com/huggingface/smolagents/pull/450
133
  # almost overwrite with https://huggingface.co/spaces/sitammeur/Gemma-llamacpp
134
  class LlamaCppModel(Model):
 
230
  agent = LlamaCppAgent(
231
  provider,
232
  system_prompt=f"{system_message}",
233
+ custom_messages_formatter=gemma_3_formatter,
234
+ #predefined_messages_formatter_type=MessagesFormatterType.GEMMA_2,
235
  debug_output=True,
236
  )
237
+ temperature = 0.5
238
  top_k=40
239
  top_p=0.95
240
+ max_tokens=2048
241
  repeat_penalty=1.1
242
  settings = provider.get_provider_default_settings()
243
  settings.temperature = temperature