Spaces:
Running on CPU Upgrade

akhaliq HF staff commited on
Commit
efff32e
·
1 Parent(s): c1081fb
Files changed (4) hide show
  1. app.py +0 -11
  2. app_huggingface.py +2 -2
  3. pyproject.toml +1 -1
  4. requirements.txt +1 -1
app.py CHANGED
@@ -1,5 +1,4 @@
1
  from app_huggingface import demo as demo_huggingface
2
- from app_openrouter import demo as demo_openrouter
3
  from utils import get_app
4
  import gradio as gr
5
 
@@ -14,22 +13,12 @@ gr.load(
14
  coder=True,
15
  provider="together"
16
  ).launch()""",
17
- "OpenRouter Coder": """
18
- import gradio as gr
19
- import ai_gradio
20
- gr.load(
21
- name='openrouter:anthropic/claude-3.7-sonnet',
22
- src=ai_gradio.registry,
23
- coder=True
24
- ).launch()""",
25
-
26
  # Add similar snippets for other providers
27
  }
28
 
29
  # Create mapping of providers to their demos
30
  PROVIDERS = {
31
  "Hugging Face": demo_huggingface,
32
- "OpenRouter Coder": demo_openrouter
33
  }
34
 
35
  # Modified get_app implementation
 
1
  from app_huggingface import demo as demo_huggingface
 
2
  from utils import get_app
3
  import gradio as gr
4
 
 
13
  coder=True,
14
  provider="together"
15
  ).launch()""",
 
 
 
 
 
 
 
 
 
16
  # Add similar snippets for other providers
17
  }
18
 
19
  # Create mapping of providers to their demos
20
  PROVIDERS = {
21
  "Hugging Face": demo_huggingface,
 
22
  }
23
 
24
  # Modified get_app implementation
app_huggingface.py CHANGED
@@ -12,10 +12,10 @@ HUGGINGFACE_MODELS_DISPLAY = [k.replace("huggingface:", "") for k in HUGGINGFACE
12
  # Create and launch the interface using get_app utility
13
  demo = get_app(
14
  models=HUGGINGFACE_MODELS_FULL, # Use the full names with prefix
15
- default_model=HUGGINGFACE_MODELS_FULL[0],
16
  dropdown_label="Select Huggingface Model",
17
  choices=HUGGINGFACE_MODELS_DISPLAY, # Display names without prefix
18
  fill_height=True,
19
  coder=True,
20
- provider="together"
21
  )
 
12
  # Create and launch the interface using get_app utility
13
  demo = get_app(
14
  models=HUGGINGFACE_MODELS_FULL, # Use the full names with prefix
15
+ default_model=HUGGINGFACE_MODELS_FULL[16],
16
  dropdown_label="Select Huggingface Model",
17
  choices=HUGGINGFACE_MODELS_DISPLAY, # Display names without prefix
18
  fill_height=True,
19
  coder=True,
20
+ provider="fireworks-ai"
21
  )
pyproject.toml CHANGED
@@ -38,7 +38,7 @@ dependencies = [
38
  "langchain>=0.3.14",
39
  "chromadb>=0.5.23",
40
  "openai>=1.55.0",
41
- "ai-gradio[crewai,deepseek,gemini,groq,hyperbolic,openai,smolagents,transformers, langchain, mistral,minimax,nvidia, qwen, openrouter, huggingface]>=0.2.50",
42
  ]
43
 
44
  [tool.uv.sources]
 
38
  "langchain>=0.3.14",
39
  "chromadb>=0.5.23",
40
  "openai>=1.55.0",
41
+ "ai-gradio[crewai,deepseek,gemini,groq,hyperbolic,openai,smolagents,transformers, langchain, mistral,minimax,nvidia, qwen, openrouter, huggingface]>=0.2.53",
42
  ]
43
 
44
  [tool.uv.sources]
requirements.txt CHANGED
@@ -2,7 +2,7 @@
2
  # uv pip compile pyproject.toml -o requirements.txt
3
  accelerate==1.2.1
4
  # via ai-gradio
5
- ai-gradio==0.2.50
6
  # via anychat (pyproject.toml)
7
  aiofiles==23.2.1
8
  # via gradio
 
2
  # uv pip compile pyproject.toml -o requirements.txt
3
  accelerate==1.2.1
4
  # via ai-gradio
5
+ ai-gradio==0.2.53
6
  # via anychat (pyproject.toml)
7
  aiofiles==23.2.1
8
  # via gradio