Update app.py
Browse files
app.py
CHANGED
@@ -19,18 +19,18 @@ For more details, please check [our post](https://huggingface.co/blog/gemma2).
|
|
19 |
|
20 |
MAX_MAX_NEW_TOKENS = 2048
|
21 |
DEFAULT_MAX_NEW_TOKENS = 1024
|
22 |
-
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "
|
23 |
|
24 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
25 |
|
26 |
-
model_id = "
|
27 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
28 |
model = AutoModelForCausalLM.from_pretrained(
|
29 |
model_id,
|
30 |
device_map="auto",
|
31 |
torch_dtype=torch.bfloat16,
|
32 |
)
|
33 |
-
model.config.sliding_window =
|
34 |
model.eval()
|
35 |
|
36 |
|
@@ -39,10 +39,7 @@ def generate(
|
|
39 |
message: str,
|
40 |
chat_history: list[dict],
|
41 |
max_new_tokens: int = 1024,
|
42 |
-
temperature: float = 0.
|
43 |
-
top_p: float = 0.9,
|
44 |
-
top_k: int = 50,
|
45 |
-
repetition_penalty: float = 1.2,
|
46 |
) -> Iterator[str]:
|
47 |
conversation = chat_history.copy()
|
48 |
conversation.append({"role": "user", "content": message})
|
@@ -116,10 +113,6 @@ demo = gr.ChatInterface(
|
|
116 |
stop_btn=None,
|
117 |
examples=[
|
118 |
["Hello there! How are you doing?"],
|
119 |
-
["Can you explain briefly to me what is the Python programming language?"],
|
120 |
-
["Explain the plot of Cinderella in a sentence."],
|
121 |
-
["How many hours does it take a man to eat a Helicopter?"],
|
122 |
-
["Write a 100-word article on 'Benefits of Open-Source in AI research'"],
|
123 |
],
|
124 |
cache_examples=False,
|
125 |
type="messages",
|
|
|
19 |
|
20 |
MAX_MAX_NEW_TOKENS = 2048
|
21 |
DEFAULT_MAX_NEW_TOKENS = 1024
|
22 |
+
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "10096"))
|
23 |
|
24 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
25 |
|
26 |
+
model_id = "silma-ai/SILMA-Kashif-2B-Instruct-v1.0"
|
27 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
28 |
model = AutoModelForCausalLM.from_pretrained(
|
29 |
model_id,
|
30 |
device_map="auto",
|
31 |
torch_dtype=torch.bfloat16,
|
32 |
)
|
33 |
+
model.config.sliding_window = 12288
|
34 |
model.eval()
|
35 |
|
36 |
|
|
|
39 |
message: str,
|
40 |
chat_history: list[dict],
|
41 |
max_new_tokens: int = 1024,
|
42 |
+
temperature: float = 0.0,
|
|
|
|
|
|
|
43 |
) -> Iterator[str]:
|
44 |
conversation = chat_history.copy()
|
45 |
conversation.append({"role": "user", "content": message})
|
|
|
113 |
stop_btn=None,
|
114 |
examples=[
|
115 |
["Hello there! How are you doing?"],
|
|
|
|
|
|
|
|
|
116 |
],
|
117 |
cache_examples=False,
|
118 |
type="messages",
|