Update app.py
Browse files
app.py
CHANGED
@@ -59,8 +59,9 @@ retriever = db.as_retriever(
|
|
59 |
model_name = "Daemontatox/CogitoZ14" # Replace with your desired model
|
60 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
61 |
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", trust_remote_code=True)
|
62 |
-
|
63 |
-
|
|
|
64 |
|
65 |
# LangChain LLM using Hugging Face Pipeline
|
66 |
llm = HuggingFacePipeline(pipeline=hf_pipeline)
|
@@ -121,4 +122,4 @@ interface = gr.Interface(
|
|
121 |
|
122 |
# Launch Gradio app
|
123 |
if __name__ == "__main__":
|
124 |
-
interface.launch()
|
|
|
59 |
model_name = "Daemontatox/CogitoZ14" # Replace with your desired model
|
60 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
61 |
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", trust_remote_code=True)
|
62 |
+
|
63 |
+
# Create Hugging Face Pipeline with the specified model and tokenizer
|
64 |
+
hf_pipeline = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
65 |
|
66 |
# LangChain LLM using Hugging Face Pipeline
|
67 |
llm = HuggingFacePipeline(pipeline=hf_pipeline)
|
|
|
122 |
|
123 |
# Launch Gradio app
|
124 |
if __name__ == "__main__":
|
125 |
+
interface.launch()
|