Luke Merrick commited on
Commit
5d1bbbd
·
1 Parent(s): 0d1661c

Update README

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -9076,7 +9076,7 @@ Key Features:
9076
 
9077
  2. Inference efficiency: Its 113m non-embedding parameters inference is fast and efficient for any scale.
9078
 
9079
- 3. Compression-friendly: Achieves high-quality retrieval with embeddings as small as 128 bytes/vector using Matryoshka Representation Learning (MRL) and quantization-aware embedding training.
9080
 
9081
  4. Long Context Support: arctic-embed-m-v2.0 builds on [GTE-multilingual-base](https://huggingface.co/Alibaba-NLP/gte-multilingual-base) which can support a context window of up to 8192 via the use of RoPE.
9082
 
 
9076
 
9077
  2. Inference efficiency: Its 113m non-embedding parameters inference is fast and efficient for any scale.
9078
 
9079
+ 3. Compression-friendly: Achieves high-quality retrieval with embeddings as small as 128 bytes/vector using Matryoshka Representation Learning (MRL) and quantization-aware embedding training. **Please note that like our v1.5 model, the MRL for this model is 256 dimensions, and high-quality 128-byte compression is achieved via 4-bit quantization (e.g. using a [`pq256x4fs` fast-scan FAISS index](https://github.com/facebookresearch/faiss/wiki/The-index-factory#encodings) or using the [example code published alongside our 1.5 model](https://github.com/Snowflake-Labs/arctic-embed/blob/main/compressed_embeddings_examples/score_arctic_embed_m_v1dot5_with_quantization.ipynb)).**
9080
 
9081
  4. Long Context Support: arctic-embed-m-v2.0 builds on [GTE-multilingual-base](https://huggingface.co/Alibaba-NLP/gte-multilingual-base) which can support a context window of up to 8192 via the use of RoPE.
9082