Update app.py
Browse files
app.py
CHANGED
@@ -106,6 +106,29 @@ hf_hub_download(
|
|
106 |
|
107 |
retriever_tool = RetrieverTool(docs_processed)
|
108 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
109 |
# based https://github.com/huggingface/smolagents/pull/450
|
110 |
# almost overwrite with https://huggingface.co/spaces/sitammeur/Gemma-llamacpp
|
111 |
class LlamaCppModel(Model):
|
@@ -207,13 +230,14 @@ class LlamaCppModel(Model):
|
|
207 |
agent = LlamaCppAgent(
|
208 |
provider,
|
209 |
system_prompt=f"{system_message}",
|
210 |
-
|
|
|
211 |
debug_output=True,
|
212 |
)
|
213 |
-
temperature = 0.
|
214 |
top_k=40
|
215 |
top_p=0.95
|
216 |
-
max_tokens=
|
217 |
repeat_penalty=1.1
|
218 |
settings = provider.get_provider_default_settings()
|
219 |
settings.temperature = temperature
|
|
|
106 |
|
107 |
retriever_tool = RetrieverTool(docs_processed)
|
108 |
|
109 |
+
# Define the prompt markers for Gemma 3
|
110 |
+
gemma_3_prompt_markers = {
|
111 |
+
Roles.system: PromptMarkers("", "\n"), # System prompt should be included within user message
|
112 |
+
Roles.user: PromptMarkers("<start_of_turn>user\n", "<end_of_turn>\n"),
|
113 |
+
Roles.assistant: PromptMarkers("<start_of_turn>model\n", "<end_of_turn>\n"),
|
114 |
+
Roles.tool: PromptMarkers("", ""), # If you need tool support
|
115 |
+
}
|
116 |
+
|
117 |
+
|
118 |
+
|
119 |
+
|
120 |
+
|
121 |
+
# Create the formatter
|
122 |
+
gemma_3_formatter = MessagesFormatter(
|
123 |
+
pre_prompt="", # No pre-prompt
|
124 |
+
prompt_markers=gemma_3_prompt_markers,
|
125 |
+
include_sys_prompt_in_first_user_message=True, # Include system prompt in first user message
|
126 |
+
default_stop_sequences=["<end_of_turn>", "<start_of_turn>"],
|
127 |
+
strip_prompt=False, # Don't strip whitespace from the prompt
|
128 |
+
bos_token="<bos>", # Beginning of sequence token for Gemma 3
|
129 |
+
eos_token="<eos>", # End of sequence token for Gemma 3
|
130 |
+
)
|
131 |
+
|
132 |
# based https://github.com/huggingface/smolagents/pull/450
|
133 |
# almost overwrite with https://huggingface.co/spaces/sitammeur/Gemma-llamacpp
|
134 |
class LlamaCppModel(Model):
|
|
|
230 |
agent = LlamaCppAgent(
|
231 |
provider,
|
232 |
system_prompt=f"{system_message}",
|
233 |
+
custom_messages_formatter=gemma_3_formatter,
|
234 |
+
#predefined_messages_formatter_type=MessagesFormatterType.GEMMA_2,
|
235 |
debug_output=True,
|
236 |
)
|
237 |
+
temperature = 0.5
|
238 |
top_k=40
|
239 |
top_p=0.95
|
240 |
+
max_tokens=2048
|
241 |
repeat_penalty=1.1
|
242 |
settings = provider.get_provider_default_settings()
|
243 |
settings.temperature = temperature
|