Updated README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,29 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- openvino
|
5 |
+
- int4
|
6 |
+
---
|
7 |
+
|
8 |
+
This is an INT4 quantized version of the `mistralai/Mistral-7B-Instruct-v0.2` model. The Python packages used in creating this model are as follows:
|
9 |
+
```
|
10 |
+
openvino==2024.3.0.dev20240528
|
11 |
+
openvino-nightly==2024.3.0.dev20240528
|
12 |
+
openvino-tokenizers==2024.3.0.0.dev20240528
|
13 |
+
optimum==1.19.2
|
14 |
+
optimum-intel==1.17.0.dev0+aefabf0
|
15 |
+
nncf==2.11.0.dev0+90a7f0d5
|
16 |
+
torch==2.3.0+cu121
|
17 |
+
transformers==4.40.2
|
18 |
+
```
|
19 |
+
This quantized model is created using the following command:
|
20 |
+
```
|
21 |
+
optimum-cli export openvino -m "mistralai/Mistral-7B-Instruct-v0.2" --task text-generation-with-past --weight-format int4 --group-size 128 --trust-remote-code ./Mistral-7B-Instruct-v0.2-ov-int4
|
22 |
+
```
|
23 |
+
For more details, run the following command from your Python environment: `optimum-cli export openvino --help`
|
24 |
+
|
25 |
+
INFO:nncf:Statistics of the bitwidth distribution:
|
26 |
+
| Num bits (N) | % all parameters (layers) | % ratio-defining parameters (layers) |
|
27 |
+
|----------------|-----------------------------|----------------------------------------|
|
28 |
+
| 8 | 23% (82 / 226) | 20% (80 / 224) |
|
29 |
+
| 4 | 77% (144 / 226) | 80% (144 / 224) |
|