Mini-GTE
Overview
This is the first model developed by QTACK and serves as a proof of concept for our distillation approach! Built upon a distillbert-based architecture, Mini-GTE is distilled from GTE and designed for efficiency without sacrificing accuracy at only 66M parameters. As a standalone sentence transformer, it ranks 2nd on the MTEB classic leaderboard in the <100M parameter category and 63rd overall which makes it a strong choice for real-time query encoding, semantic search, and similarity tasks.
Model Details
- Model Type: Sentence Transformer
- Base model: distilbert/distilbert-base-uncased
- Maximum Sequence Length: 512 tokens
- Output Dimensionality: 768 dimensions
- Similarity Function: Cosine Similarity
Usage
- Optimized for quick inference
- Great at quickly generating high quality encodings
- Easy to plug and play since it is distilled from GTE
- We want to see how you’re using our model so we’ll give you a free coffee/$10 gift card if you get on call with us and show us what you’ve built!
Getting Started
Installation
Mini-GTE is built on the Sentence Transformers framework. To install the required packages, run:
pip install -U sentence-transformers
Quick Start
Here's a quick example to get you started:
from sentence_transformers import SentenceTransformer
# Download directly from Hugging Face
model = SentenceTransformer("sentence_transformers_model_id")
# Run inference
sentences = [
'The weather is lovely today.',
"It's so sunny outside!",
'He drove to the stadium.',
]
embeddings = model.encode(sentences)
print(embeddings.shape) # Expected: [3, 768]
# Compute the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape) # Expected: [3, 3]
Training Details
- Python: 3.10.12
- Sentence Transformers: 3.3.1
- Transformers: 4.48.0.dev0
- PyTorch: 2.1.0a0+32f93b1
- Accelerate: 1.2.0
- Datasets: 2.21.0
- Tokenizers: 0.21.0
Getting Help
For any questions, suggestions, or issues, please contact the QTACK team directly through our contact page.
- Downloads last month
- 247
Model tree for prdev/mini-gte
Base model
distilbert/distilbert-base-uncasedEvaluation results
- accuracy on MTEB AmazonCounterfactualClassification (en)test set self-reported74.895
- f1 on MTEB AmazonCounterfactualClassification (en)test set self-reported68.842
- f1_weighted on MTEB AmazonCounterfactualClassification (en)test set self-reported77.182
- ap on MTEB AmazonCounterfactualClassification (en)test set self-reported37.732
- ap_weighted on MTEB AmazonCounterfactualClassification (en)test set self-reported37.732
- main_score on MTEB AmazonCounterfactualClassification (en)test set self-reported74.895
- accuracy on MTEB AmazonPolarityClassification (default)test set self-reported92.942
- f1 on MTEB AmazonPolarityClassification (default)test set self-reported92.927
- f1_weighted on MTEB AmazonPolarityClassification (default)test set self-reported92.927
- ap on MTEB AmazonPolarityClassification (default)test set self-reported89.225