zeroshot commited on
Commit
f86bf94
·
1 Parent(s): 80a660a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -1622,8 +1622,8 @@ language:
1622
  # bge-small-en-v1.5-quant
1623
 
1624
  <div>
1625
- <img src="https://huggingface.co/zeroshot/bge-small-en-v1.5-quant/resolve/main/latency.png" alt="latency" width="350" style="display:inline-block; margin-right:10px;"/>
1626
- <img src="https://huggingface.co/zeroshot/bge-small-en-v1.5-quant/resolve/main/throughput.png" alt="throughput" width="350" style="display:inline-block;"/>
1627
  </div>
1628
 
1629
  This is the quantized (INT8) ONNX variant of the [bge-small-en-v1.5](https://huggingface.co/BAAI/bge-small-en-v1.5) embeddings model created with [DeepSparse Optimum](https://github.com/neuralmagic/optimum-deepsparse) for ONNX export/inference and Neural Magic's [Sparsify](https://github.com/neuralmagic/sparsify) for one-shot quantization.
 
1622
  # bge-small-en-v1.5-quant
1623
 
1624
  <div>
1625
+ <img src="https://huggingface.co/zeroshot/bge-small-en-v1.5-quant/resolve/main/latency.png" alt="latency" width="330" style="display:inline-block; margin-right:10px;"/>
1626
+ <img src="https://huggingface.co/zeroshot/bge-small-en-v1.5-quant/resolve/main/throughput.png" alt="throughput" width="330" style="display:inline-block;"/>
1627
  </div>
1628
 
1629
  This is the quantized (INT8) ONNX variant of the [bge-small-en-v1.5](https://huggingface.co/BAAI/bge-small-en-v1.5) embeddings model created with [DeepSparse Optimum](https://github.com/neuralmagic/optimum-deepsparse) for ONNX export/inference and Neural Magic's [Sparsify](https://github.com/neuralmagic/sparsify) for one-shot quantization.