Update README.md
Browse files
README.md
CHANGED
@@ -3593,7 +3593,7 @@ model-index:
|
|
3593 |
<img src="images/gme_logo.png" alt="GME Logo" style="width: 100%; max-width: 450px;">
|
3594 |
</p>
|
3595 |
|
3596 |
-
<p align="center"><b>GME: General Multimodal
|
3597 |
|
3598 |
## GME-Qwen2-VL-2B
|
3599 |
|
@@ -3619,15 +3619,11 @@ The `GME` models support three types of input: **text**, **image**, and **image-
|
|
3619 |
## Model List
|
3620 |
| Models | Model Size | Max Seq. Length | Dimension | MTEB-en| UMRB |
|
3621 |
|:-----: | :-----: |:-----: |:-----: |:-----: | :-----: |
|
3622 |
-
|[`gme-
|
3623 |
-
|[`gme-
|
3624 |
|
3625 |
## Usage
|
3626 |
-
**Use Transformers**
|
3627 |
```
|
3628 |
-
|
3629 |
-
```
|
3630 |
-
|
3631 |
**Use with custom code**
|
3632 |
|
3633 |
```python
|
@@ -3677,8 +3673,8 @@ We validated the performance on our universal multimodal retrieval benchmark (**
|
|
3677 |
| One-Peace | 4B | 43.54 | 31.27 | 61.38 | 42.9 | 65.59 | 42.72 | 28.29 | 6.73 | 23.41 | 42.03 |
|
3678 |
| DSE | 4.2B | 48.94 | 27.92 | 40.75 | 78.21 | 52.54 | 49.62 | 35.44 | 8.36 | 40.18 | 50.63 |
|
3679 |
| E5-V | 8.4B | 52.41 | 27.36 | 46.56 | 41.22 | 47.95 | 54.13 | 32.9 | 23.17 | 7.23 | 42.48 |
|
3680 |
-
| **GME-Qwen2-VL-2B** | 2.2B | 55.93 | 29.86 | 57.36 | 87.84 |
|
3681 |
-
| **GME-Qwen2-VL-7B** | 8.3B | **58.19** | 31.89 | **61.35** | **89.92** |
|
3682 |
|
3683 |
The [MTEB Leaderboard](https://huggingface.co/spaces/mteb/leaderboard) English tab shows the text embeddings performence of our model.
|
3684 |
|
@@ -3693,7 +3689,6 @@ Due to the lack of relevant data, our models and evaluations retain one single i
|
|
3693 |
|
3694 |
We will extend to multi-image input, image-text interleaved data as well as multilingual data in the future version.
|
3695 |
|
3696 |
-
|
3697 |
## Redistribution and Use
|
3698 |
|
3699 |
We encourage and value diverse applications of GME models and continuous enhancements to the models themselves.
|
@@ -3704,12 +3699,19 @@ We encourage and value diverse applications of GME models and continuous enhance
|
|
3704 |
|
3705 |
## Cloud API Services
|
3706 |
|
3707 |
-
In addition to the open-source [GME](https://huggingface.co/collections/Alibaba-NLP/gme-models) series models, GME series models are also available as commercial API services on Alibaba Cloud.
|
3708 |
|
3709 |
- [MultiModal Embedding Models](https://help.aliyun.com/zh/model-studio/developer-reference/general-text-embedding/): The `multimodal-embedding-v1` model service is available.
|
3710 |
|
3711 |
Note that the models behind the commercial APIs are not entirely identical to the open-source models.
|
3712 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3713 |
## Citation
|
3714 |
If you find our paper or models helpful, please consider cite:
|
3715 |
|
|
|
3593 |
<img src="images/gme_logo.png" alt="GME Logo" style="width: 100%; max-width: 450px;">
|
3594 |
</p>
|
3595 |
|
3596 |
+
<p align="center"><b>GME: General Multimodal Embedding</b></p>
|
3597 |
|
3598 |
## GME-Qwen2-VL-2B
|
3599 |
|
|
|
3619 |
## Model List
|
3620 |
| Models | Model Size | Max Seq. Length | Dimension | MTEB-en| UMRB |
|
3621 |
|:-----: | :-----: |:-----: |:-----: |:-----: | :-----: |
|
3622 |
+
|[`gme-Qwen2-VL-2B`](https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-2B-Instruct) | 2.21B | 32768 | 1536 | - | 64.45 |
|
3623 |
+
|[`gme-Qwen2-VL-7B`](https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-7B-Instruct) | 8.29B | 32768 | 3584 | - | 67.44 |
|
3624 |
|
3625 |
## Usage
|
|
|
3626 |
```
|
|
|
|
|
|
|
3627 |
**Use with custom code**
|
3628 |
|
3629 |
```python
|
|
|
3673 |
| One-Peace | 4B | 43.54 | 31.27 | 61.38 | 42.9 | 65.59 | 42.72 | 28.29 | 6.73 | 23.41 | 42.03 |
|
3674 |
| DSE | 4.2B | 48.94 | 27.92 | 40.75 | 78.21 | 52.54 | 49.62 | 35.44 | 8.36 | 40.18 | 50.63 |
|
3675 |
| E5-V | 8.4B | 52.41 | 27.36 | 46.56 | 41.22 | 47.95 | 54.13 | 32.9 | 23.17 | 7.23 | 42.48 |
|
3676 |
+
| **[GME-Qwen2-VL-2B](https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-2B-Instruct)** | 2.2B | 55.93 | 29.86 | 57.36 | 87.84 | 61.93 | 76.47 | 64.58 | 37.02 | 66.47 | 64.45 |
|
3677 |
+
| **[GME-Qwen2-VL-7B](https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-7B-Instruct)** | 8.3B | **58.19** | 31.89 | **61.35** | **89.92** | **65.83** | **80.94** | **66.18** | **42.56** | **73.62** | **67.44** |
|
3678 |
|
3679 |
The [MTEB Leaderboard](https://huggingface.co/spaces/mteb/leaderboard) English tab shows the text embeddings performence of our model.
|
3680 |
|
|
|
3689 |
|
3690 |
We will extend to multi-image input, image-text interleaved data as well as multilingual data in the future version.
|
3691 |
|
|
|
3692 |
## Redistribution and Use
|
3693 |
|
3694 |
We encourage and value diverse applications of GME models and continuous enhancements to the models themselves.
|
|
|
3699 |
|
3700 |
## Cloud API Services
|
3701 |
|
3702 |
+
In addition to the open-source [GME](https://huggingface.co/collections/Alibaba-NLP/gme-models-67667e092da3491f630964d6) series models, GME series models are also available as commercial API services on Alibaba Cloud.
|
3703 |
|
3704 |
- [MultiModal Embedding Models](https://help.aliyun.com/zh/model-studio/developer-reference/general-text-embedding/): The `multimodal-embedding-v1` model service is available.
|
3705 |
|
3706 |
Note that the models behind the commercial APIs are not entirely identical to the open-source models.
|
3707 |
|
3708 |
+
## Hiring
|
3709 |
+
|
3710 |
+
We have open positions for Research Interns and Full-Time Researchers to join our team at Tongyi Lab.
|
3711 |
+
We are seeking passionate individuals with expertise in representation learning, LLM-driven information retrieval, Retrieval-Augmented Generation (RAG), and agent-based systems.
|
3712 |
+
Our team is located in the vibrant cities of Beijing and Hangzhou, offering a collaborative and dynamic work environment where you can contribute to cutting-edge advancements in artificial intelligence and machine learning.
|
3713 |
+
If you are driven by curiosity and eager to make a meaningful impact through your work, we would love to hear from you. Please submit your resume along with a brief introduction to <a href="mailto:[email protected]">[email protected]</a>.
|
3714 |
+
|
3715 |
## Citation
|
3716 |
If you find our paper or models helpful, please consider cite:
|
3717 |
|