Spaces:
Sleeping
Sleeping
Commit
·
e150690
1
Parent(s):
de90557
feat: add mistral
Browse files
app.py
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
import spaces
|
2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
3 |
|
4 |
-
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", device_map="auto")
|
5 |
-
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
|
6 |
|
7 |
# # Check if a GPU is available
|
8 |
# device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
@@ -39,6 +39,7 @@ def retriever_qa(file, query):
|
|
39 |
{"role": "user", "content": first_line}
|
40 |
]
|
41 |
|
|
|
42 |
model_inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to("cuda")
|
43 |
|
44 |
generated_ids = llm.generate(model_inputs, max_new_tokens=100, do_sample=True)
|
|
|
1 |
import spaces
|
2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
3 |
|
4 |
+
# model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", device_map="auto")
|
5 |
+
# tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
|
6 |
|
7 |
# # Check if a GPU is available
|
8 |
# device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
|
39 |
{"role": "user", "content": first_line}
|
40 |
]
|
41 |
|
42 |
+
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
|
43 |
model_inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to("cuda")
|
44 |
|
45 |
generated_ids = llm.generate(model_inputs, max_new_tokens=100, do_sample=True)
|