Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -43,9 +43,9 @@ def main():
|
|
43 |
word_count = st.sidebar.slider("Number of Words", min_value=50, max_value=1000, value=200, step=50)
|
44 |
|
45 |
if st.sidebar.button("Generate Blog"):
|
46 |
-
model_id = "mistralai/Mistral-7B-Instruct-v0.
|
47 |
-
tokenizer = AutoTokenizer.from_pretrained(model_id,
|
48 |
-
model = AutoModelForCausalLM.from_pretrained(model_id,
|
49 |
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer,max_new_tokens=1000)
|
50 |
hf = HuggingFacePipeline(pipeline=pipe)
|
51 |
chain = LLMChain(llm=hf,prompt=prompt,verbose=True)
|
|
|
43 |
word_count = st.sidebar.slider("Number of Words", min_value=50, max_value=1000, value=200, step=50)
|
44 |
|
45 |
if st.sidebar.button("Generate Blog"):
|
46 |
+
model_id = "mistralai/Mistral-7B-Instruct-v0.2"
|
47 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id,token =HF_TOKEN )
|
48 |
+
model = AutoModelForCausalLM.from_pretrained(model_id,token =HF_TOKEN )
|
49 |
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer,max_new_tokens=1000)
|
50 |
hf = HuggingFacePipeline(pipeline=pipe)
|
51 |
chain = LLMChain(llm=hf,prompt=prompt,verbose=True)
|