Dakerqi commited on
Commit
d3774a0
·
verified ·
1 Parent(s): 16a7de0

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -118,7 +118,7 @@ def model_main(args, master_port, rank, request_queue, response_queue, mp_barrie
118
 
119
  text_encoder = AutoModel.from_pretrained(
120
  "google/gemma-2-2b", torch_dtype=dtype, device_map="cuda", token=hf_token
121
- ).eval()
122
  cap_feat_dim = text_encoder.config.hidden_size
123
  if args.num_gpus > 1:
124
  raise NotImplementedError("Inference with >1 GPUs not yet supported")
 
118
 
119
  text_encoder = AutoModel.from_pretrained(
120
  "google/gemma-2-2b", torch_dtype=dtype, device_map="cuda", token=hf_token
121
+ ).eval().to("cuda")
122
  cap_feat_dim = text_encoder.config.hidden_size
123
  if args.num_gpus > 1:
124
  raise NotImplementedError("Inference with >1 GPUs not yet supported")