michaelfeil commited on
Commit
38ad519
·
1 Parent(s): 14006c8

Upload intfloat/e5-small ctranslate fp16 weights

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -2608,7 +2608,7 @@ Speedup inference while reducing memory by 2x-4x using int8 inference in C++ on
2608
 
2609
  quantized version of [intfloat/e5-small](https://huggingface.co/intfloat/e5-small)
2610
  ```bash
2611
- pip install hf-hub-ctranslate2>=2.10.0 ctranslate2>=3.16.0
2612
  ```
2613
 
2614
  ```python
@@ -2631,7 +2631,7 @@ scores = (embeddings @ embeddings.T) * 100
2631
  ```
2632
 
2633
  Checkpoint compatible to [ctranslate2>=3.16.0](https://github.com/OpenNMT/CTranslate2)
2634
- and [hf-hub-ctranslate2>=2.10.0](https://github.com/michaelfeil/hf-hub-ctranslate2)
2635
  - `compute_type=int8_float16` for `device="cuda"`
2636
  - `compute_type=int8` for `device="cpu"`
2637
 
 
2608
 
2609
  quantized version of [intfloat/e5-small](https://huggingface.co/intfloat/e5-small)
2610
  ```bash
2611
+ pip install hf-hub-ctranslate2>=3.0.0 ctranslate2>=3.16.0
2612
  ```
2613
 
2614
  ```python
 
2631
  ```
2632
 
2633
  Checkpoint compatible to [ctranslate2>=3.16.0](https://github.com/OpenNMT/CTranslate2)
2634
+ and [hf-hub-ctranslate2>=3.0.0](https://github.com/michaelfeil/hf-hub-ctranslate2)
2635
  - `compute_type=int8_float16` for `device="cuda"`
2636
  - `compute_type=int8` for `device="cpu"`
2637