use gemini for now
Browse files
app.py
CHANGED
@@ -18,7 +18,6 @@ from app_openai import demo as demo_openai
|
|
18 |
from app_paligemma import demo as demo_paligemma
|
19 |
from app_perplexity import demo as demo_perplexity
|
20 |
from app_playai import demo as demo_playai
|
21 |
-
from app_qwen import demo as demo_qwen
|
22 |
from app_replicate import demo as demo_replicate
|
23 |
from app_sambanova import demo as demo_sambanova
|
24 |
from app_showui import demo as demo_showui
|
@@ -29,7 +28,6 @@ from utils import get_app
|
|
29 |
|
30 |
# Create mapping of providers to their demos
|
31 |
PROVIDERS = {
|
32 |
-
"Qwen": demo_qwen,
|
33 |
"Gemini": demo_gemini,
|
34 |
"OpenAI Voice": demo_openai_voice,
|
35 |
"Gemini Voice": demo_gemini_voice,
|
@@ -58,7 +56,7 @@ PROVIDERS = {
|
|
58 |
"NVIDIA": demo_nvidia,
|
59 |
}
|
60 |
|
61 |
-
demo = get_app(models=list(PROVIDERS.keys()), default_model="
|
62 |
|
63 |
if __name__ == "__main__":
|
64 |
demo.queue(api_open=False).launch(show_api=False)
|
|
|
18 |
from app_paligemma import demo as demo_paligemma
|
19 |
from app_perplexity import demo as demo_perplexity
|
20 |
from app_playai import demo as demo_playai
|
|
|
21 |
from app_replicate import demo as demo_replicate
|
22 |
from app_sambanova import demo as demo_sambanova
|
23 |
from app_showui import demo as demo_showui
|
|
|
28 |
|
29 |
# Create mapping of providers to their demos
|
30 |
PROVIDERS = {
|
|
|
31 |
"Gemini": demo_gemini,
|
32 |
"OpenAI Voice": demo_openai_voice,
|
33 |
"Gemini Voice": demo_gemini_voice,
|
|
|
56 |
"NVIDIA": demo_nvidia,
|
57 |
}
|
58 |
|
59 |
+
demo = get_app(models=list(PROVIDERS.keys()), default_model="Gemini", src=PROVIDERS, dropdown_label="Select Provider")
|
60 |
|
61 |
if __name__ == "__main__":
|
62 |
demo.queue(api_open=False).launch(show_api=False)
|
utils.py
CHANGED
@@ -21,24 +21,10 @@ def get_app(
|
|
21 |
for model_name in models:
|
22 |
with gr.Column(visible=model_name == default_model) as column:
|
23 |
if isinstance(src, dict):
|
24 |
-
|
25 |
-
block = src[model_name]
|
26 |
-
else:
|
27 |
-
model_key = f"qwen:{model_name}"
|
28 |
-
block = src.get(model_key, src.get(model_name))
|
29 |
-
|
30 |
-
if isinstance(block, gr.Blocks):
|
31 |
-
block.render()
|
32 |
-
else:
|
33 |
-
# Handle the case where block is a function by passing the model name
|
34 |
-
block(name=model_name).render()
|
35 |
else:
|
36 |
-
|
37 |
-
|
38 |
-
src(model_name, accept_token).render()
|
39 |
-
else:
|
40 |
-
gr.load(name=model_name, src=src, accept_token=accept_token, **kwargs)
|
41 |
-
columns.append(column)
|
42 |
|
43 |
model.change(
|
44 |
fn=update_model,
|
@@ -51,4 +37,4 @@ def get_app(
|
|
51 |
for fn in demo.fns.values():
|
52 |
fn.api_name = False
|
53 |
|
54 |
-
return demo
|
|
|
21 |
for model_name in models:
|
22 |
with gr.Column(visible=model_name == default_model) as column:
|
23 |
if isinstance(src, dict):
|
24 |
+
src[model_name].render()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
else:
|
26 |
+
gr.load(name=model_name, src=src, accept_token=accept_token, **kwargs)
|
27 |
+
columns.append(column)
|
|
|
|
|
|
|
|
|
28 |
|
29 |
model.change(
|
30 |
fn=update_model,
|
|
|
37 |
for fn in demo.fns.values():
|
38 |
fn.api_name = False
|
39 |
|
40 |
+
return demo
|