Update README.md
Browse files
README.md
CHANGED
@@ -4,7 +4,7 @@ language:
|
|
4 |
- en
|
5 |
---
|
6 |
|
7 |
-
# codegen25-7b-multi
|
8 |
|
9 |
* Model creator: [Salesforce](https://huggingface.co/Salesforce)
|
10 |
* Original model: [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P)
|
@@ -13,19 +13,12 @@ language:
|
|
13 |
|
14 |
This is [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2024/documentation/openvino-ir-format.html) (Intermediate Representation) format with weights compressed to INT8 by [NNCF](https://github.com/openvinotoolkit/nncf).
|
15 |
|
16 |
-
## Quantization Parameters
|
17 |
-
|
18 |
-
Weight compression was performed using `nncf.compress_weights` with the following parameters:
|
19 |
-
|
20 |
-
* mode: **INT8_ASYM**
|
21 |
-
|
22 |
-
For more information on quantization, check the [OpenVINO model optimization guide](https://docs.openvino.ai/2024/openvino-workflow/model-optimization-guide/weight-compression.html).
|
23 |
|
24 |
## Compatibility
|
25 |
|
26 |
The provided OpenVINO™ IR model is compatible with:
|
27 |
|
28 |
-
* OpenVINO version 2024.
|
29 |
* Optimum Intel 1.16.0 and higher
|
30 |
|
31 |
## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
|
@@ -42,7 +35,7 @@ pip install optimum[openvino] tiktoken
|
|
42 |
from transformers import AutoTokenizer
|
43 |
from optimum.intel.openvino import OVModelForCausalLM
|
44 |
|
45 |
-
model_id = "OpenVINO/codegen25-7b-multi-
|
46 |
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
47 |
model = OVModelForCausalLM.from_pretrained(model_id)
|
48 |
text = "def hello_world():"
|
@@ -65,8 +58,8 @@ pip install openvino-genai huggingface_hub
|
|
65 |
```
|
66 |
import huggingface_hub as hf_hub
|
67 |
|
68 |
-
model_id = "OpenVINO/codegen25-7b-multi-
|
69 |
-
model_path = "codegen25-7b-multi-
|
70 |
|
71 |
hf_hub.snapshot_download(model_id, local_dir=model_path)
|
72 |
|
|
|
4 |
- en
|
5 |
---
|
6 |
|
7 |
+
# codegen25-7b-multi-fp16-ov
|
8 |
|
9 |
* Model creator: [Salesforce](https://huggingface.co/Salesforce)
|
10 |
* Original model: [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P)
|
|
|
13 |
|
14 |
This is [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2024/documentation/openvino-ir-format.html) (Intermediate Representation) format with weights compressed to INT8 by [NNCF](https://github.com/openvinotoolkit/nncf).
|
15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
## Compatibility
|
18 |
|
19 |
The provided OpenVINO™ IR model is compatible with:
|
20 |
|
21 |
+
* OpenVINO version 2024.2.0 and higher
|
22 |
* Optimum Intel 1.16.0 and higher
|
23 |
|
24 |
## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
|
|
|
35 |
from transformers import AutoTokenizer
|
36 |
from optimum.intel.openvino import OVModelForCausalLM
|
37 |
|
38 |
+
model_id = "OpenVINO/codegen25-7b-multi-fp16-ov"
|
39 |
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
40 |
model = OVModelForCausalLM.from_pretrained(model_id)
|
41 |
text = "def hello_world():"
|
|
|
58 |
```
|
59 |
import huggingface_hub as hf_hub
|
60 |
|
61 |
+
model_id = "OpenVINO/codegen25-7b-multi-fp16-ov"
|
62 |
+
model_path = "codegen25-7b-multi-fp16-ov"
|
63 |
|
64 |
hf_hub.snapshot_download(model_id, local_dir=model_path)
|
65 |
|