richardkimsm89 commited on
Commit
60d4859
·
verified ·
1 Parent(s): 90b6f18

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +62 -11
app.py CHANGED
@@ -3,23 +3,23 @@
3
  import gradio as gr
4
  from huggingface_hub import InferenceClient
5
 
6
- model = "meta-llama/Llama-3.2-3B-Instruct"
7
  client = InferenceClient(model)
8
 
9
- def fn(
10
  prompt,
11
  history,
12
  system_prompt,
13
  max_tokens,
14
  temperature,
15
  top_p,
16
- ):
17
- messages = [{"role": "system", "content": system_prompt}]
18
  history.append(messages[0])
19
 
20
- messages.append({"role": "user", "content": prompt})
21
  history.append(messages[1])
22
-
23
  stream = client.chat.completions.create(
24
  model = model,
25
  messages = history,
@@ -28,14 +28,14 @@ def fn(
28
  top_p = top_p,
29
  stream = True
30
  )
31
-
32
  chunks = []
33
  for chunk in stream:
34
  chunks.append(chunk.choices[0].delta.content or "")
35
  yield "".join(chunks)
36
 
37
- app = gr.ChatInterface(
38
- fn = fn,
39
  type = "messages",
40
  additional_inputs = [
41
  gr.Textbox(value="You are a helpful assistant.", label="System Prompt"),
@@ -47,5 +47,56 @@ app = gr.ChatInterface(
47
  description = model,
48
  )
49
 
50
- if __name__ == "__main__":
51
- app.launch()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  import gradio as gr
4
  from huggingface_hub import InferenceClient
5
 
6
+ model = "meta-llama/Llama-3.2-11B-Vision-Instruct"
7
  client = InferenceClient(model)
8
 
9
+ def fn_text(
10
  prompt,
11
  history,
12
  system_prompt,
13
  max_tokens,
14
  temperature,
15
  top_p,
16
+ ):
17
+ messages = [{"role": "system", "content": [{"type": "text", "text": system_prompt}]}]
18
  history.append(messages[0])
19
 
20
+ messages.append({"role": "user", "content": [{"type": "text", "text": prompt}]})
21
  history.append(messages[1])
22
+
23
  stream = client.chat.completions.create(
24
  model = model,
25
  messages = history,
 
28
  top_p = top_p,
29
  stream = True
30
  )
31
+
32
  chunks = []
33
  for chunk in stream:
34
  chunks.append(chunk.choices[0].delta.content or "")
35
  yield "".join(chunks)
36
 
37
+ app_text = gr.ChatInterface(
38
+ fn = fn_text,
39
  type = "messages",
40
  additional_inputs = [
41
  gr.Textbox(value="You are a helpful assistant.", label="System Prompt"),
 
47
  description = model,
48
  )
49
 
50
+ def fn_vision(
51
+ prompt,
52
+ image_url,
53
+ #system_prompt,
54
+ max_tokens,
55
+ temperature,
56
+ top_p,
57
+ ):
58
+ messages = [{"role": "user", "content": [{"type": "text", "text": prompt}]}]
59
+
60
+ if image_url:
61
+ messages[0]["content"].append({"type": "image_url", "image_url": {"url": image_url}})
62
+
63
+ stream = client.chat.completions.create(
64
+ model = model,
65
+ messages = messages,
66
+ max_tokens = max_tokens,
67
+ temperature = temperature,
68
+ top_p = top_p,
69
+ stream = True
70
+ )
71
+
72
+ chunks = []
73
+ for chunk in stream:
74
+ chunks.append(chunk.choices[0].delta.content or "")
75
+ yield "".join(chunks)
76
+
77
+ app_vision = gr.Interface(
78
+ fn = fn_vision,
79
+ inputs = [
80
+ gr.Textbox(label="Prompt"),
81
+ gr.Textbox(label="Image URL")
82
+ ],
83
+ outputs = [
84
+ gr.Textbox(label="Output")
85
+ ],
86
+ additional_inputs = [
87
+ #gr.Textbox(value="You are a helpful assistant.", label="System Prompt"),
88
+ gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max Tokens"),
89
+ gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
90
+ gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-P"),
91
+ ],
92
+ title = "Meta Llama",
93
+ description = model,
94
+ )
95
+
96
+ app = gr.TabbedInterface(
97
+ [app_text, app_vision],
98
+ ["Text", "Vision"]
99
+ ).launch()
100
+
101
+ #if __name__ == "__main__":
102
+ # app.launch()