kamran-r123 commited on
Commit
ecadc7d
·
verified ·
1 Parent(s): f6819c7

Update main.py

Browse files
Files changed (1) hide show
  1. main.py +4 -1
main.py CHANGED
@@ -6,12 +6,13 @@ import prompt_style
6
 
7
 
8
  model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3"
9
- client = InferenceClient(model_id)
10
 
11
  class Item(BaseModel):
12
  prompt: str
13
  history: list
14
  system_prompt: str
 
15
  temperature: float = 0.6
16
  max_new_tokens: int = 1024
17
  top_p: float = 0.95
@@ -44,6 +45,8 @@ def generate(item: Item):
44
  seed=item.seed,
45
  )
46
 
 
 
47
  formatted_prompt = format_prompt(item)
48
  stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
49
  output = ""
 
6
 
7
 
8
  model_id = "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3"
9
+ client = None
10
 
11
  class Item(BaseModel):
12
  prompt: str
13
  history: list
14
  system_prompt: str
15
+ token:str
16
  temperature: float = 0.6
17
  max_new_tokens: int = 1024
18
  top_p: float = 0.95
 
45
  seed=item.seed,
46
  )
47
 
48
+ if not client:
49
+ client = InferenceClient(token=item.token, model=model_id)
50
  formatted_prompt = format_prompt(item)
51
  stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
52
  output = ""