karimouda commited on
Commit
8174431
·
verified ·
1 Parent(s): 4468ddf

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -11
app.py CHANGED
@@ -19,18 +19,18 @@ For more details, please check [our post](https://huggingface.co/blog/gemma2).
19
 
20
  MAX_MAX_NEW_TOKENS = 2048
21
  DEFAULT_MAX_NEW_TOKENS = 1024
22
- MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
23
 
24
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
25
 
26
- model_id = "google/gemma-2-9b-it"
27
  tokenizer = AutoTokenizer.from_pretrained(model_id)
28
  model = AutoModelForCausalLM.from_pretrained(
29
  model_id,
30
  device_map="auto",
31
  torch_dtype=torch.bfloat16,
32
  )
33
- model.config.sliding_window = 4096
34
  model.eval()
35
 
36
 
@@ -39,10 +39,7 @@ def generate(
39
  message: str,
40
  chat_history: list[dict],
41
  max_new_tokens: int = 1024,
42
- temperature: float = 0.6,
43
- top_p: float = 0.9,
44
- top_k: int = 50,
45
- repetition_penalty: float = 1.2,
46
  ) -> Iterator[str]:
47
  conversation = chat_history.copy()
48
  conversation.append({"role": "user", "content": message})
@@ -116,10 +113,6 @@ demo = gr.ChatInterface(
116
  stop_btn=None,
117
  examples=[
118
  ["Hello there! How are you doing?"],
119
- ["Can you explain briefly to me what is the Python programming language?"],
120
- ["Explain the plot of Cinderella in a sentence."],
121
- ["How many hours does it take a man to eat a Helicopter?"],
122
- ["Write a 100-word article on 'Benefits of Open-Source in AI research'"],
123
  ],
124
  cache_examples=False,
125
  type="messages",
 
19
 
20
  MAX_MAX_NEW_TOKENS = 2048
21
  DEFAULT_MAX_NEW_TOKENS = 1024
22
+ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "10096"))
23
 
24
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
25
 
26
+ model_id = "silma-ai/SILMA-Kashif-2B-Instruct-v1.0"
27
  tokenizer = AutoTokenizer.from_pretrained(model_id)
28
  model = AutoModelForCausalLM.from_pretrained(
29
  model_id,
30
  device_map="auto",
31
  torch_dtype=torch.bfloat16,
32
  )
33
+ model.config.sliding_window = 12288
34
  model.eval()
35
 
36
 
 
39
  message: str,
40
  chat_history: list[dict],
41
  max_new_tokens: int = 1024,
42
+ temperature: float = 0.0,
 
 
 
43
  ) -> Iterator[str]:
44
  conversation = chat_history.copy()
45
  conversation.append({"role": "user", "content": message})
 
113
  stop_btn=None,
114
  examples=[
115
  ["Hello there! How are you doing?"],
 
 
 
 
116
  ],
117
  cache_examples=False,
118
  type="messages",