davanstrien HF Staff commited on
Commit
af9c7a0
·
1 Parent(s): 6311bc5

Adjust GPU memory utilization default and add max model length argument in generate-responses.py

Browse files
Files changed (1) hide show
  1. generate-responses.py +8 -2
generate-responses.py CHANGED
@@ -397,8 +397,13 @@ Examples:
397
  parser.add_argument(
398
  "--gpu-memory-utilization",
399
  type=float,
400
- default=0.95,
401
- help="GPU memory utilization factor (default: 0.95)",
 
 
 
 
 
402
  )
403
  parser.add_argument(
404
  "--tensor-parallel-size",
@@ -426,6 +431,7 @@ Examples:
426
  max_tokens=args.max_tokens,
427
  repetition_penalty=args.repetition_penalty,
428
  gpu_memory_utilization=args.gpu_memory_utilization,
 
429
  tensor_parallel_size=args.tensor_parallel_size,
430
  hf_token=args.hf_token,
431
  )
 
397
  parser.add_argument(
398
  "--gpu-memory-utilization",
399
  type=float,
400
+ default=0.90,
401
+ help="GPU memory utilization factor (default: 0.90)",
402
+ )
403
+ parser.add_argument(
404
+ "--max-model-len",
405
+ type=int,
406
+ help="Maximum model context length (default: model's default)",
407
  )
408
  parser.add_argument(
409
  "--tensor-parallel-size",
 
431
  max_tokens=args.max_tokens,
432
  repetition_penalty=args.repetition_penalty,
433
  gpu_memory_utilization=args.gpu_memory_utilization,
434
+ max_model_len=args.max_model_len,
435
  tensor_parallel_size=args.tensor_parallel_size,
436
  hf_token=args.hf_token,
437
  )