Spaces:
Runtime error
Runtime error
update code
Browse files
app.py
CHANGED
@@ -38,14 +38,14 @@ def how_to_load(language):
|
|
38 |
def load_model(values, language):
|
39 |
model = values["model"]
|
40 |
if not model:
|
41 |
-
text = f"""No model available for {language.capitalize()}. If you trained a model on this language, let us know in\
|
42 |
in the [Community tab](https://huggingface.co/spaces/loubnabnl/the-stack-bot/discussions) to feature your model!\n\
|
43 |
You can also train your own model on The Stack using the instructions below 🚀"""
|
44 |
st.write(text)
|
45 |
if st.button("Fine-tune your own model", key=4):
|
46 |
st.write("Code available at [GitHub link] + add preview")
|
47 |
else:
|
48 |
-
text = f"""{model} is a model
|
49 |
code = f"""
|
50 |
```python
|
51 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
@@ -106,7 +106,7 @@ if st.session_state["Models trained on dataset"]:
|
|
106 |
load_model(languages[selected_language], selected_language)
|
107 |
|
108 |
if languages[selected_language]["model"] and languages[selected_language]["gradio_demo"]:
|
109 |
-
st.write(f"Here's a demo to try
|
110 |
gen_prompt = st.text_area(
|
111 |
"Generate code with prompt:",
|
112 |
value="# Implement a function to print hello world",
|
|
|
38 |
def load_model(values, language):
|
39 |
model = values["model"]
|
40 |
if not model:
|
41 |
+
text = f"""No model is available for {language.capitalize()}. If you trained a model on this language, let us know in\
|
42 |
in the [Community tab](https://huggingface.co/spaces/loubnabnl/the-stack-bot/discussions) to feature your model!\n\
|
43 |
You can also train your own model on The Stack using the instructions below 🚀"""
|
44 |
st.write(text)
|
45 |
if st.button("Fine-tune your own model", key=4):
|
46 |
st.write("Code available at [GitHub link] + add preview")
|
47 |
else:
|
48 |
+
text = f"""[{model}](hf.co/{model}) is a model trained on the {language.capitalize()} subset of The Stack. Here's how to use it:"""
|
49 |
code = f"""
|
50 |
```python
|
51 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
106 |
load_model(languages[selected_language], selected_language)
|
107 |
|
108 |
if languages[selected_language]["model"] and languages[selected_language]["gradio_demo"]:
|
109 |
+
st.write(f"Here's a demo to try it, for more flexibilty you can use the original [Gradio demo]({languages[selected_language]['gradio_demo']}).")
|
110 |
gen_prompt = st.text_area(
|
111 |
"Generate code with prompt:",
|
112 |
value="# Implement a function to print hello world",
|