finalf0 commited on
Commit
04b63b5
·
1 Parent(s): 6f1c207

Update readme, example code use bfloat16 to reduce GPU memory usage

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -124,7 +124,8 @@ import torch
124
  from PIL import Image
125
  from transformers import AutoModel, AutoTokenizer
126
 
127
- model = AutoModel.from_pretrained('openbmb/MiniCPM-V', trust_remote_code=True).to(dtype=torch.bfloat16)
 
128
  tokenizer = AutoTokenizer.from_pretrained('openbmb/MiniCPM-V', trust_remote_code=True)
129
  model.eval().cuda()
130
 
 
124
  from PIL import Image
125
  from transformers import AutoModel, AutoTokenizer
126
 
127
+ model = AutoModel.from_pretrained('openbmb/MiniCPM-V', trust_remote_code=True)
128
+ model = model.to(dtype=torch.bfloat16)
129
  tokenizer = AutoTokenizer.from_pretrained('openbmb/MiniCPM-V', trust_remote_code=True)
130
  model.eval().cuda()
131