Sam Heutmaker
commited on
Commit
·
0efb203
1
Parent(s):
d42fbaf
update readme
Browse files
README.md
CHANGED
@@ -39,7 +39,7 @@ model-index:
|
|
39 |
|
40 |
## Model Description
|
41 |
|
42 |
-
**ClipTagger-12b** is a 12-billion parameter vision-language model (VLM) designed for video understanding at massive scale. Developed by [Inference.net](https://inference.net) in collaboration with [Grass](https://grass.io), this model was created to meet the demanding requirements of trillion-scale video frame captioning workloads
|
43 |
|
44 |
**ClipTagger-12b exceeds or matches the performance of GPT-4.1 and Claude 4 Sonnet, while costing 15x less per generation.**
|
45 |
|
@@ -234,4 +234,4 @@ Contact us at [[email protected]](mailto:[email protected]) for a free
|
|
234 |
|
235 |
## License
|
236 |
|
237 |
-
This model is released under the Apache-2.0 license, allowing for commercial use and modification with proper attribution.
|
|
|
39 |
|
40 |
## Model Description
|
41 |
|
42 |
+
**ClipTagger-12b** is a 12-billion parameter vision-language model (VLM) designed for video understanding at massive scale. Developed by [Inference.net](https://inference.net) in collaboration with [Grass](https://grass.io), this model was created to meet the demanding requirements of trillion-scale video frame captioning workloads.
|
43 |
|
44 |
**ClipTagger-12b exceeds or matches the performance of GPT-4.1 and Claude 4 Sonnet, while costing 15x less per generation.**
|
45 |
|
|
|
234 |
|
235 |
## License
|
236 |
|
237 |
+
This model is released under the Apache-2.0 license, allowing for commercial use and modification with proper attribution.
|