InspirationYF commited on
Commit
e150690
·
1 Parent(s): de90557

feat: add mistral

Browse files
Files changed (1) hide show
  1. app.py +3 -2
app.py CHANGED
@@ -1,8 +1,8 @@
1
  import spaces
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
- model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", device_map="auto")
5
- tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
6
 
7
  # # Check if a GPU is available
8
  # device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
@@ -39,6 +39,7 @@ def retriever_qa(file, query):
39
  {"role": "user", "content": first_line}
40
  ]
41
 
 
42
  model_inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to("cuda")
43
 
44
  generated_ids = llm.generate(model_inputs, max_new_tokens=100, do_sample=True)
 
1
  import spaces
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
+ # model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", device_map="auto")
5
+ # tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
6
 
7
  # # Check if a GPU is available
8
  # device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
 
39
  {"role": "user", "content": first_line}
40
  ]
41
 
42
+ tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
43
  model_inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to("cuda")
44
 
45
  generated_ids = llm.generate(model_inputs, max_new_tokens=100, do_sample=True)