Update README.md
Browse files
README.md
CHANGED
@@ -9,9 +9,11 @@ tags:
|
|
9 |
|
10 |

|
11 |
|
12 |
-
Ivy\-VL is a lightweight multimodal model with only 3B parameters. It accepts both image and text inputs to generate text outputs.
|
13 |
|
14 |
-
Thanks to its lightweight design, it can be deployed on edge devices such as AI glasses and smartphones, offering low memory usage and high speed while maintaining strong performance on multimodal tasks.
|
|
|
|
|
15 |
|
16 |
# Model Summary:
|
17 |
|
@@ -103,7 +105,7 @@ print(text_outputs)
|
|
103 |
|
104 |
```plaintext
|
105 |
@misc{ivy2024ivy-vl,
|
106 |
-
title={
|
107 |
url={https://huggingface.co/AI-Safeguard/Ivy-VL},
|
108 |
author={Ivy Zhang,Jenny N,Theresa Yu and David Qiu},
|
109 |
month={December},
|
|
|
9 |
|
10 |

|
11 |
|
12 |
+
[Ivy\-VL] is a lightweight multimodal model with only 3B parameters. It accepts both image and text inputs to generate text outputs.
|
13 |
|
14 |
+
Thanks to its lightweight design, it can be deployed on edge devices such as AI glasses and smartphones, offering low memory usage and high speed while maintaining strong performance on multimodal tasks.
|
15 |
+
|
16 |
+
The model is built upon the [Qwen/Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct) language model, with [google/siglip-so400m-patch14-384](https://huggingface.co/google/siglip-so400m-patch14-384) serving as the vision encoder.
|
17 |
|
18 |
# Model Summary:
|
19 |
|
|
|
105 |
|
106 |
```plaintext
|
107 |
@misc{ivy2024ivy-vl,
|
108 |
+
title={Ivy-VL:Compact Vision-Language Models Achieving SOTA with Optimal Data},
|
109 |
url={https://huggingface.co/AI-Safeguard/Ivy-VL},
|
110 |
author={Ivy Zhang,Jenny N,Theresa Yu and David Qiu},
|
111 |
month={December},
|