vermen commited on
Commit
b8c33f8
·
verified ·
1 Parent(s): 274c4d3

Changing response length

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -16,7 +16,7 @@ llm = LlamaCPP(
16
  # optionally, you can set the path to a pre-downloaded model instead of model_url
17
  model_path=None,
18
  temperature=0.01,
19
- max_new_tokens=1024,
20
  # llama2 has a context window of 4096 tokens, but we set it lower to allow for some wiggle room
21
  context_window=3900,
22
  # kwargs to pass to __call__()
 
16
  # optionally, you can set the path to a pre-downloaded model instead of model_url
17
  model_path=None,
18
  temperature=0.01,
19
+ max_new_tokens=256, # could be larger but requires more time
20
  # llama2 has a context window of 4096 tokens, but we set it lower to allow for some wiggle room
21
  context_window=3900,
22
  # kwargs to pass to __call__()