timxiaohangt commited on
Commit
6c30c0d
·
1 Parent(s): f3cf99c
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -7,13 +7,13 @@ from pydantic import BaseModel
7
  from transformers import AutoTokenizer, AutoModelForCausalLM, LlamaTokenizer
8
 
9
  os.environ["CUDA_DEVICE_ORDER"]="PCI_BUS_ID"
10
- os.environ['CUDA_VISIBLE_DEVICES'] = "0"
11
 
12
  MODEL_NAME = "RegularizedSelfPlay/sppo_forward1reverse5-0.1-PromptABC-Mistral-7B-Instruct-SPPO-Iter3" # Example: "meta-llama/Llama-2-7b-chat-hf"
13
  HF_TOKEN = os.getenv("HF_API_TOKEN")
14
 
15
  # Load model and tokenizer
16
- tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", token=HF_TOKEN)
17
  llm = LLM(
18
  model=MODEL_NAME,
19
  # revision="1296dc8fd9b21e6424c9c305c06db9ae60c03ace",
 
7
  from transformers import AutoTokenizer, AutoModelForCausalLM, LlamaTokenizer
8
 
9
  os.environ["CUDA_DEVICE_ORDER"]="PCI_BUS_ID"
10
+ os.environ['CUDA_VISIBLE_DEVICES'] = "2"
11
 
12
  MODEL_NAME = "RegularizedSelfPlay/sppo_forward1reverse5-0.1-PromptABC-Mistral-7B-Instruct-SPPO-Iter3" # Example: "meta-llama/Llama-2-7b-chat-hf"
13
  HF_TOKEN = os.getenv("HF_API_TOKEN")
14
 
15
  # Load model and tokenizer
16
+ tokenizer = LlamaTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", token=HF_TOKEN)
17
  llm = LLM(
18
  model=MODEL_NAME,
19
  # revision="1296dc8fd9b21e6424c9c305c06db9ae60c03ace",