Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,6 @@ This model is an int4 model with group_size 128 and symmetric quantization of [d
|
|
11 |
## How to Use
|
12 |
### INT4 Inference
|
13 |
```python
|
14 |
-
from auto_round import AutoRoundConfig ##must import for auto-round format
|
15 |
import requests
|
16 |
import torch
|
17 |
from PIL import Image
|
@@ -21,7 +20,7 @@ from deepseek_vl2.models import DeepseekVLV2Processor, DeepseekVLV2ForCausalLM
|
|
21 |
|
22 |
|
23 |
# specify the path to the model
|
24 |
-
model_path = "OPEA/deepseek-vl2-int4-sym-inc"
|
25 |
vl_chat_processor: DeepseekVLV2Processor = DeepseekVLV2Processor.from_pretrained(model_path)
|
26 |
tokenizer = vl_chat_processor.tokenizer
|
27 |
|
|
|
11 |
## How to Use
|
12 |
### INT4 Inference
|
13 |
```python
|
|
|
14 |
import requests
|
15 |
import torch
|
16 |
from PIL import Image
|
|
|
20 |
|
21 |
|
22 |
# specify the path to the model
|
23 |
+
model_path = "OPEA/deepseek-vl2-int4-sym-gptq-inc"
|
24 |
vl_chat_processor: DeepseekVLV2Processor = DeepseekVLV2Processor.from_pretrained(model_path)
|
25 |
tokenizer = vl_chat_processor.tokenizer
|
26 |
|