Update README.md
Browse files
README.md
CHANGED
@@ -235,7 +235,7 @@ lm_eval \
|
|
235 |
## Inference Performance
|
236 |
|
237 |
|
238 |
-
This model achieves up to
|
239 |
The following performance benchmarks were conducted with [vLLM](https://docs.vllm.ai/en/latest/) version 0.7.2, and [GuideLLM](https://github.com/neuralmagic/guidellm).
|
240 |
|
241 |
<details>
|
@@ -387,22 +387,22 @@ The following performance benchmarks were conducted with [vLLM](https://docs.vll
|
|
387 |
<tr>
|
388 |
<td>neuralmagic/Qwen2-VL-72B-Instruct-quantized.w8a8</td>
|
389 |
<td>1.84</td>
|
390 |
-
<td>
|
391 |
-
<td>
|
392 |
-
<td>
|
393 |
-
<td>
|
394 |
-
<td>
|
395 |
-
<td>
|
396 |
</tr>
|
397 |
<tr>
|
398 |
<td>neuralmagic/Qwen2-VL-72B-Instruct-quantized.w4a16</td>
|
399 |
<td>2.73</td>
|
400 |
-
<td>
|
401 |
-
<td>
|
402 |
-
<td>
|
403 |
-
<td>
|
404 |
-
<td>
|
405 |
-
<td>
|
406 |
</tr>
|
407 |
<tr>
|
408 |
<th rowspan="3" valign="top">H100x4</td>
|
@@ -418,22 +418,22 @@ The following performance benchmarks were conducted with [vLLM](https://docs.vll
|
|
418 |
<tr>
|
419 |
<td>neuralmagic/Qwen2-VL-72B-Instruct-FP8-Dynamic</td>
|
420 |
<td>1.70</td>
|
421 |
-
<td>
|
422 |
-
<td>
|
423 |
-
<td>
|
424 |
-
<td>
|
425 |
-
<td>4
|
426 |
-
<td>
|
427 |
</tr>
|
428 |
<tr>
|
429 |
<td>neuralmagic/Qwen2-VL-72B-Instruct-quantized.w4a16</td>
|
430 |
<td>2.35</td>
|
431 |
-
<td>
|
432 |
-
<td>
|
433 |
-
<td>
|
434 |
-
<td>
|
435 |
-
<td>
|
436 |
-
<td>
|
437 |
</tr>
|
438 |
</tbody>
|
439 |
</table>
|
|
|
235 |
## Inference Performance
|
236 |
|
237 |
|
238 |
+
This model achieves up to 1.84x speedup in single-stream deployment and up to 1.85x speedup in multi-stream asynchronous deployment, depending on hardware and use-case scenario.
|
239 |
The following performance benchmarks were conducted with [vLLM](https://docs.vllm.ai/en/latest/) version 0.7.2, and [GuideLLM](https://github.com/neuralmagic/guidellm).
|
240 |
|
241 |
<details>
|
|
|
387 |
<tr>
|
388 |
<td>neuralmagic/Qwen2-VL-72B-Instruct-quantized.w8a8</td>
|
389 |
<td>1.84</td>
|
390 |
+
<td>0.6</td>
|
391 |
+
<td>293</td>
|
392 |
+
<td>2.0</td>
|
393 |
+
<td>1021</td>
|
394 |
+
<td>2.3</td>
|
395 |
+
<td>1135</td>
|
396 |
</tr>
|
397 |
<tr>
|
398 |
<td>neuralmagic/Qwen2-VL-72B-Instruct-quantized.w4a16</td>
|
399 |
<td>2.73</td>
|
400 |
+
<td>0.6</td>
|
401 |
+
<td>314</td>
|
402 |
+
<td>3.2</td>
|
403 |
+
<td>1591</td>
|
404 |
+
<td>4.0</td>
|
405 |
+
<td>2019</td>
|
406 |
</tr>
|
407 |
<tr>
|
408 |
<th rowspan="3" valign="top">H100x4</td>
|
|
|
418 |
<tr>
|
419 |
<td>neuralmagic/Qwen2-VL-72B-Instruct-FP8-Dynamic</td>
|
420 |
<td>1.70</td>
|
421 |
+
<td>0.8</td>
|
422 |
+
<td>236</td>
|
423 |
+
<td>2.2</td>
|
424 |
+
<td>623</td>
|
425 |
+
<td>2.4</td>
|
426 |
+
<td>669</td>
|
427 |
</tr>
|
428 |
<tr>
|
429 |
<td>neuralmagic/Qwen2-VL-72B-Instruct-quantized.w4a16</td>
|
430 |
<td>2.35</td>
|
431 |
+
<td>1.3</td>
|
432 |
+
<td>350</td>
|
433 |
+
<td>3.3</td>
|
434 |
+
<td>910</td>
|
435 |
+
<td>3.6</td>
|
436 |
+
<td>994</td>
|
437 |
</tr>
|
438 |
</tbody>
|
439 |
</table>
|