Update README.md
Browse files
README.md
CHANGED
|
@@ -39,12 +39,8 @@ The embedding module **roberta.embeddings.token_type_embeddings** falls back to
|
|
| 39 |
|
| 40 |
### Test result
|
| 41 |
|
| 42 |
-
- Batch size = 8
|
| 43 |
-
- [Amazon Web Services](https://aws.amazon.com/) c6i.xlarge (Intel ICE Lake: 4 vCPUs, 8g Memory) instance.
|
| 44 |
-
|
| 45 |
| |INT8|FP32|
|
| 46 |
|---|:---:|:---:|
|
| 47 |
-
| **Throughput (samples/sec)** |25.737|13.171|
|
| 48 |
| **Accuracy (eval-f1)** |0.9247|0.9138|
|
| 49 |
| **Model size (MB)** |121|476|
|
| 50 |
|
|
@@ -56,7 +52,3 @@ int8_model = OptimizedModel.from_pretrained(
|
|
| 56 |
'Intel/roberta-base-mrpc-int8-static',
|
| 57 |
)
|
| 58 |
```
|
| 59 |
-
|
| 60 |
-
Notes:
|
| 61 |
-
- The INT8 model has better performance than the FP32 model when the CPU is fully occupied. Otherwise, there will be the illusion that INT8 is inferior to FP32.
|
| 62 |
-
|
|
|
|
| 39 |
|
| 40 |
### Test result
|
| 41 |
|
|
|
|
|
|
|
|
|
|
| 42 |
| |INT8|FP32|
|
| 43 |
|---|:---:|:---:|
|
|
|
|
| 44 |
| **Accuracy (eval-f1)** |0.9247|0.9138|
|
| 45 |
| **Model size (MB)** |121|476|
|
| 46 |
|
|
|
|
| 52 |
'Intel/roberta-base-mrpc-int8-static',
|
| 53 |
)
|
| 54 |
```
|
|
|
|
|
|
|
|
|
|
|
|