katuni4ka commited on
Commit
84c16ba
·
verified ·
1 Parent(s): c8220ca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -12
README.md CHANGED
@@ -4,7 +4,7 @@ language:
4
  - en
5
  ---
6
 
7
- # codegen25-7b-multi
8
 
9
  * Model creator: [Salesforce](https://huggingface.co/Salesforce)
10
  * Original model: [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P)
@@ -13,19 +13,12 @@ language:
13
 
14
  This is [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2024/documentation/openvino-ir-format.html) (Intermediate Representation) format with weights compressed to INT8 by [NNCF](https://github.com/openvinotoolkit/nncf).
15
 
16
- ## Quantization Parameters
17
-
18
- Weight compression was performed using `nncf.compress_weights` with the following parameters:
19
-
20
- * mode: **INT8_ASYM**
21
-
22
- For more information on quantization, check the [OpenVINO model optimization guide](https://docs.openvino.ai/2024/openvino-workflow/model-optimization-guide/weight-compression.html).
23
 
24
  ## Compatibility
25
 
26
  The provided OpenVINO™ IR model is compatible with:
27
 
28
- * OpenVINO version 2024.1.0 and higher
29
  * Optimum Intel 1.16.0 and higher
30
 
31
  ## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
@@ -42,7 +35,7 @@ pip install optimum[openvino] tiktoken
42
  from transformers import AutoTokenizer
43
  from optimum.intel.openvino import OVModelForCausalLM
44
 
45
- model_id = "OpenVINO/codegen25-7b-multi-int8-ov"
46
  tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
47
  model = OVModelForCausalLM.from_pretrained(model_id)
48
  text = "def hello_world():"
@@ -65,8 +58,8 @@ pip install openvino-genai huggingface_hub
65
  ```
66
  import huggingface_hub as hf_hub
67
 
68
- model_id = "OpenVINO/codegen25-7b-multi-int8-ov"
69
- model_path = "codegen25-7b-multi-int8-ov"
70
 
71
  hf_hub.snapshot_download(model_id, local_dir=model_path)
72
 
 
4
  - en
5
  ---
6
 
7
+ # codegen25-7b-multi-fp16-ov
8
 
9
  * Model creator: [Salesforce](https://huggingface.co/Salesforce)
10
  * Original model: [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P)
 
13
 
14
  This is [CodeGen2.5-7B-multi](https://huggingface.co/Salesforce/codegen25-7b-multi_P) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2024/documentation/openvino-ir-format.html) (Intermediate Representation) format with weights compressed to INT8 by [NNCF](https://github.com/openvinotoolkit/nncf).
15
 
 
 
 
 
 
 
 
16
 
17
  ## Compatibility
18
 
19
  The provided OpenVINO™ IR model is compatible with:
20
 
21
+ * OpenVINO version 2024.2.0 and higher
22
  * Optimum Intel 1.16.0 and higher
23
 
24
  ## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
 
35
  from transformers import AutoTokenizer
36
  from optimum.intel.openvino import OVModelForCausalLM
37
 
38
+ model_id = "OpenVINO/codegen25-7b-multi-fp16-ov"
39
  tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
40
  model = OVModelForCausalLM.from_pretrained(model_id)
41
  text = "def hello_world():"
 
58
  ```
59
  import huggingface_hub as hf_hub
60
 
61
+ model_id = "OpenVINO/codegen25-7b-multi-fp16-ov"
62
+ model_path = "codegen25-7b-multi-fp16-ov"
63
 
64
  hf_hub.snapshot_download(model_id, local_dir=model_path)
65