Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Commit
·
78e9bc6
1
Parent(s):
b9392b7
New API models
Browse files- README.md +38 -0
- external_models_results.json +158 -3
- model_list.txt +38 -0
- update_models_in_readme.py +6 -2
README.md
CHANGED
@@ -113,6 +113,7 @@ models:
|
|
113 |
- EleutherAI/gpt-neo-125m
|
114 |
- EleutherAI/gpt-neo-2.7B
|
115 |
- EleutherAI/gpt-neox-20b
|
|
|
116 |
- EleutherAI/pythia-12b
|
117 |
- EleutherAI/pythia-14m
|
118 |
- EleutherAI/pythia-160m
|
@@ -135,8 +136,11 @@ models:
|
|
135 |
- EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
|
136 |
- EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
|
137 |
- EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
|
|
|
138 |
- EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
|
139 |
- Eurdem/Defne-llama3.1-8B
|
|
|
|
|
140 |
- FuseAI/FuseChat-7B-VaRM
|
141 |
- FuseAI/OpenChat-3.5-7B-Solar
|
142 |
- GeneZC/MiniChat-1.5-3B
|
@@ -201,6 +205,7 @@ models:
|
|
201 |
- Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
|
202 |
- Magpie-Align/MagpieLM-8B-Chat-v0.1
|
203 |
- Magpie-Align/MagpieLM-8B-SFT-v0.1
|
|
|
204 |
- MagusCorp/legislinho
|
205 |
- MarinaraSpaghetti/NemoReRemix-12B
|
206 |
- MaziyarPanahi/Calme-4x7B-MoE-v0.1
|
@@ -212,6 +217,10 @@ models:
|
|
212 |
- MaziyarPanahi/Mistral-7B-Instruct-v0.3
|
213 |
- MaziyarPanahi/Mistral-7B-v0.3
|
214 |
- MaziyarPanahi/Topxtral-4x7B-v0.1
|
|
|
|
|
|
|
|
|
215 |
- MaziyarPanahi/calme-2.7-qwen2-7b
|
216 |
- MulaBR/Mula-4x160-v0.1
|
217 |
- MulaBR/Mula-8x160-v0.1
|
@@ -221,6 +230,8 @@ models:
|
|
221 |
- NTQAI/Nxcode-CQ-7B-orpo
|
222 |
- Nexusflow/Starling-LM-7B-beta
|
223 |
- Nos-PT/Carvalho_pt-gl-1.3B
|
|
|
|
|
224 |
- NotAiLOL/Yi-1.5-dolphin-9B
|
225 |
- NousResearch/Hermes-2-Pro-Llama-3-8B
|
226 |
- NousResearch/Hermes-2-Theta-Llama-3-8B
|
@@ -293,6 +304,7 @@ models:
|
|
293 |
- Qwen/Qwen2.5-1.5B-Instruct
|
294 |
- Qwen/Qwen2.5-14B
|
295 |
- Qwen/Qwen2.5-14B-Instruct
|
|
|
296 |
- Qwen/Qwen2.5-32B
|
297 |
- Qwen/Qwen2.5-32B-Instruct
|
298 |
- Qwen/Qwen2.5-3B
|
@@ -307,8 +319,13 @@ models:
|
|
307 |
- Qwen/Qwen2.5-Coder-32B-Instruct
|
308 |
- Qwen/Qwen2.5-Coder-7B
|
309 |
- Qwen/Qwen2.5-Coder-7B-Instruct
|
|
|
|
|
|
|
|
|
310 |
- RLHFlow/LLaMA3-iterative-DPO-final
|
311 |
- Ramikan-BR/Qwen2-0.5B-v5
|
|
|
312 |
- Ramikan-BR/TiamaPY-v30
|
313 |
- Ramikan-BR/TiamaPY-v31
|
314 |
- Ramikan-BR/tinyllama-coder-py-4bit-v10
|
@@ -320,7 +337,9 @@ models:
|
|
320 |
- RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
|
321 |
- Sakalti/Oxyge1-33B
|
322 |
- SakanaAI/DiscoPOP-zephyr-7b-gemma
|
|
|
323 |
- Sao10K/L3-8B-Niitama-v1
|
|
|
324 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
|
325 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
|
326 |
- SeaLLMs/SeaLLM-7B-v2
|
@@ -347,6 +366,7 @@ models:
|
|
347 |
- TheBloke/wizardLM-7B-HF
|
348 |
- TheBloke/zephyr-7B-beta-GPTQ
|
349 |
- TheDrummer/Big-Tiger-Gemma-27B-v1
|
|
|
350 |
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
|
351 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
352 |
- TinyLlama/TinyLlama_v1.1
|
@@ -368,6 +388,7 @@ models:
|
|
368 |
- VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
|
369 |
- VAGOsolutions/SauerkrautLM-Gemma-7b
|
370 |
- VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
|
|
|
371 |
- VAGOsolutions/SauerkrautLM-Qwen-32b
|
372 |
- ValiantLabs/Llama3.1-8B-Cobalt
|
373 |
- ValiantLabs/Llama3.1-8B-Fireplace2
|
@@ -397,9 +418,12 @@ models:
|
|
397 |
- Xwin-LM/Xwin-LM-7B-V0.2
|
398 |
- abacusai/Liberated-Qwen1.5-14B
|
399 |
- abacusai/Llama-3-Smaug-8B
|
|
|
400 |
- abacusai/Smaug-34B-v0.1
|
401 |
- abacusai/Smaug-72B-v0.1
|
402 |
- abacusai/Smaug-Llama-3-70B-Instruct
|
|
|
|
|
403 |
- abhishek/autotrain-llama3-orpo-v2
|
404 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese
|
405 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
|
@@ -445,6 +469,7 @@ models:
|
|
445 |
- arcee-ai/Arcee-Spark
|
446 |
- arcee-ai/Llama-3.1-SuperNova-Lite
|
447 |
- arcee-ai/SuperNova-Medius
|
|
|
448 |
- arcee-ai/Virtuoso-Small
|
449 |
- argilla/CapybaraHermes-2.5-Mistral-7B
|
450 |
- argilla/notus-7b-v1
|
@@ -472,6 +497,7 @@ models:
|
|
472 |
- botbot-ai/CabraLlama3-8b
|
473 |
- botbot-ai/CabraMistral-v3-7b-32k
|
474 |
- botbot-ai/CabraMixtral-8x7b
|
|
|
475 |
- byroneverson/Mistral-Small-Instruct-2409-abliterated
|
476 |
- byroneverson/Yi-1.5-9B-Chat-16K-abliterated
|
477 |
- byroneverson/Yi-1.5-9B-Chat-abliterated
|
@@ -489,6 +515,10 @@ models:
|
|
489 |
- chujiezheng/Mistral7B-PairRM-SPPO-ExPO
|
490 |
- chujiezheng/Smaug-34B-v0.1-ExPO
|
491 |
- cnmoro/Mistral-7B-Portuguese
|
|
|
|
|
|
|
|
|
492 |
- cognitivecomputations/WestLake-7B-v2-laser
|
493 |
- cognitivecomputations/WizardLM-13B-Uncensored
|
494 |
- cognitivecomputations/WizardLM-30B-Uncensored
|
@@ -510,6 +540,7 @@ models:
|
|
510 |
- cognitivess/bella-1-8b
|
511 |
- cosmicvalor/mistral-orthogonalized
|
512 |
- croissantllm/CroissantLLMBase
|
|
|
513 |
- deepseek-ai/DeepSeek-R1-Distill-Llama-8B
|
514 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
515 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
|
@@ -573,6 +604,7 @@ models:
|
|
573 |
- freewheelin/free-solar-evo-v0.1
|
574 |
- freewheelin/free-solar-evo-v0.11
|
575 |
- freewheelin/free-solar-evo-v0.13
|
|
|
576 |
- ghost-x/ghost-8b-beta
|
577 |
- ghost-x/ghost-8b-beta-1608
|
578 |
- google/gemma-1.1-2b-it
|
@@ -733,6 +765,7 @@ models:
|
|
733 |
- microsoft/phi-1_5
|
734 |
- microsoft/phi-2
|
735 |
- microsoft/phi-4
|
|
|
736 |
- migtissera/Tess-M-v1.3
|
737 |
- migtissera/Tess-v2.5-Gemma-2-27B-alpha
|
738 |
- migtissera/Tess-v2.5-Phi-3-medium-128k-14B
|
@@ -766,12 +799,14 @@ models:
|
|
766 |
- mlabonne/NeuralDaredevil-8B-abliterated
|
767 |
- mlabonne/NeuralMonarch-7B
|
768 |
- monilouise/opt125M_portuguese
|
|
|
769 |
- mosaicml/mpt-30b
|
770 |
- mosaicml/mpt-7b
|
771 |
- mosaicml/mpt-7b-8k
|
772 |
- natong19/Mistral-Nemo-Instruct-2407-abliterated
|
773 |
- natong19/Qwen2-7B-Instruct-abliterated
|
774 |
- nbeerbower/gemma2-gutenberg-27B
|
|
|
775 |
- nicholasKluge/Aira-2-portuguese-124M
|
776 |
- nicholasKluge/Aira-2-portuguese-1B7
|
777 |
- nicholasKluge/Aira-2-portuguese-560M
|
@@ -855,6 +890,7 @@ models:
|
|
855 |
- princeton-nlp/Mistral-7B-Instruct-SimPO
|
856 |
- princeton-nlp/gemma-2-9b-it-DPO
|
857 |
- princeton-nlp/gemma-2-9b-it-SimPO
|
|
|
858 |
- prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
|
859 |
- projecte-aina/FLOR-1.3B
|
860 |
- projecte-aina/FLOR-6.3B
|
@@ -933,11 +969,13 @@ models:
|
|
933 |
- shadowml/BeagSake-7B
|
934 |
- shadowml/Mixolar-4x7b
|
935 |
- sometimesanotion/Lamarck-14B-v0.6
|
|
|
936 |
- sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
|
937 |
- speakleash/Bielik-11B-v2
|
938 |
- speakleash/Bielik-11B-v2.0-Instruct
|
939 |
- speakleash/Bielik-11B-v2.1-Instruct
|
940 |
- speakleash/Bielik-11B-v2.2-Instruct
|
|
|
941 |
- ssmits/Falcon2-5.5B-Portuguese
|
942 |
- ssmits/Falcon2-5.5B-multilingual
|
943 |
- stabilityai/stablelm-2-12b
|
|
|
113 |
- EleutherAI/gpt-neo-125m
|
114 |
- EleutherAI/gpt-neo-2.7B
|
115 |
- EleutherAI/gpt-neox-20b
|
116 |
+
- EleutherAI/polyglot-ko-12.8b
|
117 |
- EleutherAI/pythia-12b
|
118 |
- EleutherAI/pythia-14m
|
119 |
- EleutherAI/pythia-160m
|
|
|
136 |
- EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
|
137 |
- EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
|
138 |
- EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
|
139 |
+
- EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos
|
140 |
- EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
|
141 |
- Eurdem/Defne-llama3.1-8B
|
142 |
+
- FluxiIA/Tucaninho
|
143 |
+
- FluxiIA/Tucaninho_dpo
|
144 |
- FuseAI/FuseChat-7B-VaRM
|
145 |
- FuseAI/OpenChat-3.5-7B-Solar
|
146 |
- GeneZC/MiniChat-1.5-3B
|
|
|
205 |
- Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
|
206 |
- Magpie-Align/MagpieLM-8B-Chat-v0.1
|
207 |
- Magpie-Align/MagpieLM-8B-SFT-v0.1
|
208 |
+
- MagusCorp/grpo_lora_enem_llama3_7b
|
209 |
- MagusCorp/legislinho
|
210 |
- MarinaraSpaghetti/NemoReRemix-12B
|
211 |
- MaziyarPanahi/Calme-4x7B-MoE-v0.1
|
|
|
217 |
- MaziyarPanahi/Mistral-7B-Instruct-v0.3
|
218 |
- MaziyarPanahi/Mistral-7B-v0.3
|
219 |
- MaziyarPanahi/Topxtral-4x7B-v0.1
|
220 |
+
- MaziyarPanahi/calme-2.1-qwen2-7b
|
221 |
+
- MaziyarPanahi/calme-2.2-qwen2-7b
|
222 |
+
- MaziyarPanahi/calme-2.3-qwen2-7b
|
223 |
+
- MaziyarPanahi/calme-2.4-qwen2-7b
|
224 |
- MaziyarPanahi/calme-2.7-qwen2-7b
|
225 |
- MulaBR/Mula-4x160-v0.1
|
226 |
- MulaBR/Mula-8x160-v0.1
|
|
|
230 |
- NTQAI/Nxcode-CQ-7B-orpo
|
231 |
- Nexusflow/Starling-LM-7B-beta
|
232 |
- Nos-PT/Carvalho_pt-gl-1.3B
|
233 |
+
- Nos-PT/Llama-Carvalho-PT
|
234 |
+
- Nos-PT/Llama-Carvalho-PT-GL
|
235 |
- NotAiLOL/Yi-1.5-dolphin-9B
|
236 |
- NousResearch/Hermes-2-Pro-Llama-3-8B
|
237 |
- NousResearch/Hermes-2-Theta-Llama-3-8B
|
|
|
304 |
- Qwen/Qwen2.5-1.5B-Instruct
|
305 |
- Qwen/Qwen2.5-14B
|
306 |
- Qwen/Qwen2.5-14B-Instruct
|
307 |
+
- Qwen/Qwen2.5-14B-Instruct-1M
|
308 |
- Qwen/Qwen2.5-32B
|
309 |
- Qwen/Qwen2.5-32B-Instruct
|
310 |
- Qwen/Qwen2.5-3B
|
|
|
319 |
- Qwen/Qwen2.5-Coder-32B-Instruct
|
320 |
- Qwen/Qwen2.5-Coder-7B
|
321 |
- Qwen/Qwen2.5-Coder-7B-Instruct
|
322 |
+
- Qwen/Qwen2.5-VL-32B-Instruct
|
323 |
+
- Qwen/Qwen2.5-VL-72B-Instruct
|
324 |
+
- RDson/WomboCombo-R1-Coder-14B-Preview
|
325 |
+
- RLHFlow/ArmoRM-Llama3-8B-v0.1
|
326 |
- RLHFlow/LLaMA3-iterative-DPO-final
|
327 |
- Ramikan-BR/Qwen2-0.5B-v5
|
328 |
+
- Ramikan-BR/Qwen2-0.5B-v9
|
329 |
- Ramikan-BR/TiamaPY-v30
|
330 |
- Ramikan-BR/TiamaPY-v31
|
331 |
- Ramikan-BR/tinyllama-coder-py-4bit-v10
|
|
|
337 |
- RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
|
338 |
- Sakalti/Oxyge1-33B
|
339 |
- SakanaAI/DiscoPOP-zephyr-7b-gemma
|
340 |
+
- Salesforce/SFR-Iterative-DPO-LLaMA-3-8B-R
|
341 |
- Sao10K/L3-8B-Niitama-v1
|
342 |
+
- Sao10K/L3.1-8B-Niitama-v1.1
|
343 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
|
344 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
|
345 |
- SeaLLMs/SeaLLM-7B-v2
|
|
|
366 |
- TheBloke/wizardLM-7B-HF
|
367 |
- TheBloke/zephyr-7B-beta-GPTQ
|
368 |
- TheDrummer/Big-Tiger-Gemma-27B-v1
|
369 |
+
- TheDrummer/Gemmasutra-9B-v1
|
370 |
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
|
371 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
372 |
- TinyLlama/TinyLlama_v1.1
|
|
|
388 |
- VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
|
389 |
- VAGOsolutions/SauerkrautLM-Gemma-7b
|
390 |
- VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
|
391 |
+
- VAGOsolutions/SauerkrautLM-Phi-3-medium
|
392 |
- VAGOsolutions/SauerkrautLM-Qwen-32b
|
393 |
- ValiantLabs/Llama3.1-8B-Cobalt
|
394 |
- ValiantLabs/Llama3.1-8B-Fireplace2
|
|
|
418 |
- Xwin-LM/Xwin-LM-7B-V0.2
|
419 |
- abacusai/Liberated-Qwen1.5-14B
|
420 |
- abacusai/Llama-3-Smaug-8B
|
421 |
+
- abacusai/Slerp-CM-mist-dpo
|
422 |
- abacusai/Smaug-34B-v0.1
|
423 |
- abacusai/Smaug-72B-v0.1
|
424 |
- abacusai/Smaug-Llama-3-70B-Instruct
|
425 |
+
- abacusai/bigstral-12b-32k
|
426 |
+
- abacusai/bigyi-15b
|
427 |
- abhishek/autotrain-llama3-orpo-v2
|
428 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese
|
429 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
|
|
|
469 |
- arcee-ai/Arcee-Spark
|
470 |
- arcee-ai/Llama-3.1-SuperNova-Lite
|
471 |
- arcee-ai/SuperNova-Medius
|
472 |
+
- arcee-ai/Virtuoso-Lite
|
473 |
- arcee-ai/Virtuoso-Small
|
474 |
- argilla/CapybaraHermes-2.5-Mistral-7B
|
475 |
- argilla/notus-7b-v1
|
|
|
497 |
- botbot-ai/CabraLlama3-8b
|
498 |
- botbot-ai/CabraMistral-v3-7b-32k
|
499 |
- botbot-ai/CabraMixtral-8x7b
|
500 |
+
- bunnycore/HyperLlama-3.1-8B
|
501 |
- byroneverson/Mistral-Small-Instruct-2409-abliterated
|
502 |
- byroneverson/Yi-1.5-9B-Chat-16K-abliterated
|
503 |
- byroneverson/Yi-1.5-9B-Chat-abliterated
|
|
|
515 |
- chujiezheng/Mistral7B-PairRM-SPPO-ExPO
|
516 |
- chujiezheng/Smaug-34B-v0.1-ExPO
|
517 |
- cnmoro/Mistral-7B-Portuguese
|
518 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-Hybrid-Reasoning
|
519 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-v1
|
520 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-v2
|
521 |
+
- cognitivecomputations/Dolphin3.0-R1-Mistral-24B
|
522 |
- cognitivecomputations/WestLake-7B-v2-laser
|
523 |
- cognitivecomputations/WizardLM-13B-Uncensored
|
524 |
- cognitivecomputations/WizardLM-30B-Uncensored
|
|
|
540 |
- cognitivess/bella-1-8b
|
541 |
- cosmicvalor/mistral-orthogonalized
|
542 |
- croissantllm/CroissantLLMBase
|
543 |
+
- deepseek-ai/DeepSeek-R1-Distill-Llama-70B
|
544 |
- deepseek-ai/DeepSeek-R1-Distill-Llama-8B
|
545 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
546 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
|
|
|
604 |
- freewheelin/free-solar-evo-v0.1
|
605 |
- freewheelin/free-solar-evo-v0.11
|
606 |
- freewheelin/free-solar-evo-v0.13
|
607 |
+
- gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
|
608 |
- ghost-x/ghost-8b-beta
|
609 |
- ghost-x/ghost-8b-beta-1608
|
610 |
- google/gemma-1.1-2b-it
|
|
|
765 |
- microsoft/phi-1_5
|
766 |
- microsoft/phi-2
|
767 |
- microsoft/phi-4
|
768 |
+
- migtissera/Tess-3-Mistral-Nemo-12B
|
769 |
- migtissera/Tess-M-v1.3
|
770 |
- migtissera/Tess-v2.5-Gemma-2-27B-alpha
|
771 |
- migtissera/Tess-v2.5-Phi-3-medium-128k-14B
|
|
|
799 |
- mlabonne/NeuralDaredevil-8B-abliterated
|
800 |
- mlabonne/NeuralMonarch-7B
|
801 |
- monilouise/opt125M_portuguese
|
802 |
+
- monsterapi/Llama-3_1-8B-Instruct-orca-ORPO
|
803 |
- mosaicml/mpt-30b
|
804 |
- mosaicml/mpt-7b
|
805 |
- mosaicml/mpt-7b-8k
|
806 |
- natong19/Mistral-Nemo-Instruct-2407-abliterated
|
807 |
- natong19/Qwen2-7B-Instruct-abliterated
|
808 |
- nbeerbower/gemma2-gutenberg-27B
|
809 |
+
- nbeerbower/mistral-nemo-wissenschaft-12B
|
810 |
- nicholasKluge/Aira-2-portuguese-124M
|
811 |
- nicholasKluge/Aira-2-portuguese-1B7
|
812 |
- nicholasKluge/Aira-2-portuguese-560M
|
|
|
890 |
- princeton-nlp/Mistral-7B-Instruct-SimPO
|
891 |
- princeton-nlp/gemma-2-9b-it-DPO
|
892 |
- princeton-nlp/gemma-2-9b-it-SimPO
|
893 |
+
- prithivMLmods/Megatron-Opus-14B-Exp
|
894 |
- prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
|
895 |
- projecte-aina/FLOR-1.3B
|
896 |
- projecte-aina/FLOR-6.3B
|
|
|
969 |
- shadowml/BeagSake-7B
|
970 |
- shadowml/Mixolar-4x7b
|
971 |
- sometimesanotion/Lamarck-14B-v0.6
|
972 |
+
- sometimesanotion/Lamarck-14B-v0.7
|
973 |
- sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
|
974 |
- speakleash/Bielik-11B-v2
|
975 |
- speakleash/Bielik-11B-v2.0-Instruct
|
976 |
- speakleash/Bielik-11B-v2.1-Instruct
|
977 |
- speakleash/Bielik-11B-v2.2-Instruct
|
978 |
+
- speakleash/Bielik-11B-v2.3-Instruct
|
979 |
- ssmits/Falcon2-5.5B-Portuguese
|
980 |
- ssmits/Falcon2-5.5B-multilingual
|
981 |
- stabilityai/stablelm-2-12b
|
external_models_results.json
CHANGED
@@ -176,8 +176,8 @@
|
|
176 |
"result_metrics_npm": 0.6834036936130392
|
177 |
},
|
178 |
{
|
179 |
-
"model": "gemini-1.5-flash",
|
180 |
-
"name": "Gemini 1.5 Flash",
|
181 |
"link": "https://cloud.google.com/vertex-ai",
|
182 |
"date": "2024-08-09",
|
183 |
"status": "full",
|
@@ -222,7 +222,7 @@
|
|
222 |
{
|
223 |
"model": "nemotron-4-340b-instruct",
|
224 |
"name": "nvidia/Nemotron-4-340B-Instruct (Nvidia API)",
|
225 |
-
"link": "https://
|
226 |
"date": "2024-06-30",
|
227 |
"status": "full",
|
228 |
"main_language": "English",
|
@@ -333,6 +333,51 @@
|
|
333 |
"result_metrics_average": 0.7928134532766066,
|
334 |
"result_metrics_npm": 0.6915070359785283
|
335 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
336 |
{
|
337 |
"model": "gemini-2.0-flash-001",
|
338 |
"name": "Gemini 2.0 Flash (001)",
|
@@ -376,5 +421,115 @@
|
|
376 |
},
|
377 |
"result_metrics_average": 0.8056048352614735,
|
378 |
"result_metrics_npm": 0.6986042497176748
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
379 |
}
|
380 |
]
|
|
|
176 |
"result_metrics_npm": 0.6834036936130392
|
177 |
},
|
178 |
{
|
179 |
+
"model": "gemini-1.5-flash-001",
|
180 |
+
"name": "Gemini 1.5 Flash (001)",
|
181 |
"link": "https://cloud.google.com/vertex-ai",
|
182 |
"date": "2024-08-09",
|
183 |
"status": "full",
|
|
|
222 |
{
|
223 |
"model": "nemotron-4-340b-instruct",
|
224 |
"name": "nvidia/Nemotron-4-340B-Instruct (Nvidia API)",
|
225 |
+
"link": "https://huggingface.co/nvidia/Nemotron-4-340B-Instruct",
|
226 |
"date": "2024-06-30",
|
227 |
"status": "full",
|
228 |
"main_language": "English",
|
|
|
333 |
"result_metrics_average": 0.7928134532766066,
|
334 |
"result_metrics_npm": 0.6915070359785283
|
335 |
},
|
336 |
+
{
|
337 |
+
"model": "gemini-1.5-flash-002",
|
338 |
+
"name": "Gemini 1.5 Flash (002)",
|
339 |
+
"link": "https://cloud.google.com/vertex-ai",
|
340 |
+
"date": "2025-04-03",
|
341 |
+
"status": "full",
|
342 |
+
"main_language": "English",
|
343 |
+
"model_type": "proprietary",
|
344 |
+
"result_metrics": {
|
345 |
+
"enem_challenge": 0.8327501749475158,
|
346 |
+
"bluex": 0.760778859527121,
|
347 |
+
"oab_exams": 0.6369020501138952,
|
348 |
+
"assin2_sts": 0.8380176734291938,
|
349 |
+
"assin2_rte": 0.941176117215237,
|
350 |
+
"faquad_nli": 0.8360786822325283,
|
351 |
+
"hatebr_offensive": 0.9046145161133335,
|
352 |
+
"portuguese_hate_speech": 0.7406414313684444,
|
353 |
+
"tweetsentbr": 0.6997509880131249
|
354 |
+
},
|
355 |
+
"result_metrics_average": 0.7989678325511549,
|
356 |
+
"result_metrics_npm": 0.6979777100000177
|
357 |
+
},
|
358 |
+
{
|
359 |
+
"model": "gemini-1.5-flash-8b-001",
|
360 |
+
"name": "Gemini 1.5 Flash 8B (001)",
|
361 |
+
"link": "https://aistudio.google.com",
|
362 |
+
"date": "2025-04-03",
|
363 |
+
"status": "full",
|
364 |
+
"main_language": "English",
|
365 |
+
"model_type": "proprietary",
|
366 |
+
"params": 8.0,
|
367 |
+
"result_metrics": {
|
368 |
+
"enem_challenge": 0.7641707487753674,
|
369 |
+
"bluex": 0.6467315716272601,
|
370 |
+
"oab_exams": 0.5603644646924829,
|
371 |
+
"assin2_sts": 0.7638946799836569,
|
372 |
+
"assin2_rte": 0.9329452628161146,
|
373 |
+
"faquad_nli": 0.7937022965448601,
|
374 |
+
"hatebr_offensive": 0.850497640901663,
|
375 |
+
"portuguese_hate_speech": 0.7391317606010173,
|
376 |
+
"tweetsentbr": 0.7376684798923661
|
377 |
+
},
|
378 |
+
"result_metrics_average": 0.7543452117594209,
|
379 |
+
"result_metrics_npm": 0.6359642422837162
|
380 |
+
},
|
381 |
{
|
382 |
"model": "gemini-2.0-flash-001",
|
383 |
"name": "Gemini 2.0 Flash (001)",
|
|
|
421 |
},
|
422 |
"result_metrics_average": 0.8056048352614735,
|
423 |
"result_metrics_npm": 0.6986042497176748
|
424 |
+
},
|
425 |
+
{
|
426 |
+
"model": "gemini-2.5-pro-exp-03-25",
|
427 |
+
"name": "Gemini 2.5 Pro Experimental (0325)",
|
428 |
+
"link": "https://aistudio.google.com",
|
429 |
+
"date": "2025-04-03",
|
430 |
+
"status": "full",
|
431 |
+
"main_language": "English",
|
432 |
+
"model_type": "proprietary",
|
433 |
+
"result_metrics": {
|
434 |
+
"enem_challenge": 0.9769069279216235,
|
435 |
+
"bluex": 0.9499304589707928,
|
436 |
+
"oab_exams": 0.9216400911161731,
|
437 |
+
"assin2_sts": 0.837785744915033,
|
438 |
+
"assin2_rte": 0.9415510158830285,
|
439 |
+
"faquad_nli": 0.8738735797309651,
|
440 |
+
"hatebr_offensive": 0.9248478168290788,
|
441 |
+
"portuguese_hate_speech": 0.7336133105156697,
|
442 |
+
"tweetsentbr": 0.7928002469993594
|
443 |
+
},
|
444 |
+
"result_metrics_average": 0.8836610214313025,
|
445 |
+
"result_metrics_npm": 0.8134610556797854
|
446 |
+
},
|
447 |
+
{
|
448 |
+
"model": "qwen2-5-vl-72b-instruct",
|
449 |
+
"name": "Qwen/Qwen2.5-VL-72B-Instruct (API)",
|
450 |
+
"link": "https://huggingface.co/Qwen/Qwen2.5-VL-72B-Instruct",
|
451 |
+
"date": "2025-04-03",
|
452 |
+
"status": "full",
|
453 |
+
"main_language": "English",
|
454 |
+
"model_type": "chat",
|
455 |
+
"result_metrics": {
|
456 |
+
"enem_challenge": 0.8600419874037789,
|
457 |
+
"bluex": 0.8052851182197497,
|
458 |
+
"oab_exams": 0.6888382687927107,
|
459 |
+
"assin2_sts": 0.7595538567467497,
|
460 |
+
"assin2_rte": 0.9472975104201871,
|
461 |
+
"faquad_nli": 0.8447190882122586,
|
462 |
+
"hatebr_offensive": 0.8810695094657859,
|
463 |
+
"portuguese_hate_speech": 0.769596419318135,
|
464 |
+
"tweetsentbr": 0.5644757075411895
|
465 |
+
},
|
466 |
+
"result_metrics_average": 0.7912086073467273,
|
467 |
+
"result_metrics_npm": 0.6888261361422966
|
468 |
+
},
|
469 |
+
{
|
470 |
+
"model": "qwen2-5-72b-instruct",
|
471 |
+
"name": "Qwen/Qwen2.5-72B-Instruct (API)",
|
472 |
+
"link": "https://huggingface.co/Qwen/Qwen2.5-72B-Instruct",
|
473 |
+
"date": "2025-04-03",
|
474 |
+
"status": "full",
|
475 |
+
"main_language": "English",
|
476 |
+
"model_type": "chat",
|
477 |
+
"result_metrics": {
|
478 |
+
"enem_challenge": 0.8432470258922323,
|
479 |
+
"bluex": 0.780250347705146,
|
480 |
+
"oab_exams": 0.675626423690205,
|
481 |
+
"assin2_sts": 0.8230708844558656,
|
482 |
+
"assin2_rte": 0.9509720145268106,
|
483 |
+
"faquad_nli": 0.8194444444444444,
|
484 |
+
"hatebr_offensive": 0.8810033427242816,
|
485 |
+
"portuguese_hate_speech": 0.7601866578782712,
|
486 |
+
"tweetsentbr": 0.7620172222071487
|
487 |
+
},
|
488 |
+
"result_metrics_average": 0.8106464848360451,
|
489 |
+
"result_metrics_npm": 0.7142994872542282
|
490 |
+
},
|
491 |
+
{
|
492 |
+
"model": "qwen2-5-vl-32b-instruct",
|
493 |
+
"name": "Qwen/Qwen2.5-VL-32B-Instruct (API)",
|
494 |
+
"link": "https://huggingface.co/Qwen/Qwen2.5-VL-32B-Instruct",
|
495 |
+
"date": "2025-04-03",
|
496 |
+
"status": "full",
|
497 |
+
"main_language": "English",
|
498 |
+
"model_type": "chat",
|
499 |
+
"result_metrics": {
|
500 |
+
"enem_challenge": 0.8600419874037789,
|
501 |
+
"bluex": 0.8052851182197497,
|
502 |
+
"oab_exams": 0.6888382687927107,
|
503 |
+
"assin2_sts": 0.7780549055529008,
|
504 |
+
"assin2_rte": 0.9472975104201871,
|
505 |
+
"faquad_nli": 0.8447190882122586,
|
506 |
+
"hatebr_offensive": 0.8810695094657859,
|
507 |
+
"portuguese_hate_speech": 0.769596419318135,
|
508 |
+
"tweetsentbr": 0.7027408707999051
|
509 |
+
},
|
510 |
+
"result_metrics_average": 0.8086270753539346,
|
511 |
+
"result_metrics_npm": 0.7137431116807307
|
512 |
+
},
|
513 |
+
{
|
514 |
+
"model": "qwen-turbo-2024-11-01",
|
515 |
+
"name": "Qwen-Turbo (2024-11-01)",
|
516 |
+
"link": "https://www.alibabacloud.com/en/product/modelstudio",
|
517 |
+
"date": "2025-04-03",
|
518 |
+
"status": "full",
|
519 |
+
"main_language": "English",
|
520 |
+
"model_type": "proprietary",
|
521 |
+
"result_metrics": {
|
522 |
+
"enem_challenge": 0.7795661301609517,
|
523 |
+
"bluex": 0.7079276773296245,
|
524 |
+
"oab_exams": 0.6091116173120729,
|
525 |
+
"assin2_sts": 0.7640477700456898,
|
526 |
+
"assin2_rte": 0.9260451969385788,
|
527 |
+
"faquad_nli": 0.8128063725490196,
|
528 |
+
"hatebr_offensive": 0.8567933277676292,
|
529 |
+
"portuguese_hate_speech": 0.7239183383094245,
|
530 |
+
"tweetsentbr": 0.7038360447972195
|
531 |
+
},
|
532 |
+
"result_metrics_average": 0.7648947194678011,
|
533 |
+
"result_metrics_npm": 0.6490441260447987
|
534 |
}
|
535 |
]
|
model_list.txt
CHANGED
@@ -84,6 +84,7 @@
|
|
84 |
- EleutherAI/gpt-neo-125m
|
85 |
- EleutherAI/gpt-neo-2.7B
|
86 |
- EleutherAI/gpt-neox-20b
|
|
|
87 |
- EleutherAI/pythia-12b
|
88 |
- EleutherAI/pythia-14m
|
89 |
- EleutherAI/pythia-160m
|
@@ -106,8 +107,11 @@
|
|
106 |
- EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
|
107 |
- EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
|
108 |
- EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
|
|
|
109 |
- EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
|
110 |
- Eurdem/Defne-llama3.1-8B
|
|
|
|
|
111 |
- FuseAI/FuseChat-7B-VaRM
|
112 |
- FuseAI/OpenChat-3.5-7B-Solar
|
113 |
- GeneZC/MiniChat-1.5-3B
|
@@ -172,6 +176,7 @@
|
|
172 |
- Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
|
173 |
- Magpie-Align/MagpieLM-8B-Chat-v0.1
|
174 |
- Magpie-Align/MagpieLM-8B-SFT-v0.1
|
|
|
175 |
- MagusCorp/legislinho
|
176 |
- MarinaraSpaghetti/NemoReRemix-12B
|
177 |
- MaziyarPanahi/Calme-4x7B-MoE-v0.1
|
@@ -183,6 +188,10 @@
|
|
183 |
- MaziyarPanahi/Mistral-7B-Instruct-v0.3
|
184 |
- MaziyarPanahi/Mistral-7B-v0.3
|
185 |
- MaziyarPanahi/Topxtral-4x7B-v0.1
|
|
|
|
|
|
|
|
|
186 |
- MaziyarPanahi/calme-2.7-qwen2-7b
|
187 |
- MulaBR/Mula-4x160-v0.1
|
188 |
- MulaBR/Mula-8x160-v0.1
|
@@ -192,6 +201,8 @@
|
|
192 |
- NTQAI/Nxcode-CQ-7B-orpo
|
193 |
- Nexusflow/Starling-LM-7B-beta
|
194 |
- Nos-PT/Carvalho_pt-gl-1.3B
|
|
|
|
|
195 |
- NotAiLOL/Yi-1.5-dolphin-9B
|
196 |
- NousResearch/Hermes-2-Pro-Llama-3-8B
|
197 |
- NousResearch/Hermes-2-Theta-Llama-3-8B
|
@@ -264,6 +275,7 @@
|
|
264 |
- Qwen/Qwen2.5-1.5B-Instruct
|
265 |
- Qwen/Qwen2.5-14B
|
266 |
- Qwen/Qwen2.5-14B-Instruct
|
|
|
267 |
- Qwen/Qwen2.5-32B
|
268 |
- Qwen/Qwen2.5-32B-Instruct
|
269 |
- Qwen/Qwen2.5-3B
|
@@ -278,8 +290,13 @@
|
|
278 |
- Qwen/Qwen2.5-Coder-32B-Instruct
|
279 |
- Qwen/Qwen2.5-Coder-7B
|
280 |
- Qwen/Qwen2.5-Coder-7B-Instruct
|
|
|
|
|
|
|
|
|
281 |
- RLHFlow/LLaMA3-iterative-DPO-final
|
282 |
- Ramikan-BR/Qwen2-0.5B-v5
|
|
|
283 |
- Ramikan-BR/TiamaPY-v30
|
284 |
- Ramikan-BR/TiamaPY-v31
|
285 |
- Ramikan-BR/tinyllama-coder-py-4bit-v10
|
@@ -291,7 +308,9 @@
|
|
291 |
- RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
|
292 |
- Sakalti/Oxyge1-33B
|
293 |
- SakanaAI/DiscoPOP-zephyr-7b-gemma
|
|
|
294 |
- Sao10K/L3-8B-Niitama-v1
|
|
|
295 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
|
296 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
|
297 |
- SeaLLMs/SeaLLM-7B-v2
|
@@ -318,6 +337,7 @@
|
|
318 |
- TheBloke/wizardLM-7B-HF
|
319 |
- TheBloke/zephyr-7B-beta-GPTQ
|
320 |
- TheDrummer/Big-Tiger-Gemma-27B-v1
|
|
|
321 |
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
|
322 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
323 |
- TinyLlama/TinyLlama_v1.1
|
@@ -339,6 +359,7 @@
|
|
339 |
- VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
|
340 |
- VAGOsolutions/SauerkrautLM-Gemma-7b
|
341 |
- VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
|
|
|
342 |
- VAGOsolutions/SauerkrautLM-Qwen-32b
|
343 |
- ValiantLabs/Llama3.1-8B-Cobalt
|
344 |
- ValiantLabs/Llama3.1-8B-Fireplace2
|
@@ -368,9 +389,12 @@
|
|
368 |
- Xwin-LM/Xwin-LM-7B-V0.2
|
369 |
- abacusai/Liberated-Qwen1.5-14B
|
370 |
- abacusai/Llama-3-Smaug-8B
|
|
|
371 |
- abacusai/Smaug-34B-v0.1
|
372 |
- abacusai/Smaug-72B-v0.1
|
373 |
- abacusai/Smaug-Llama-3-70B-Instruct
|
|
|
|
|
374 |
- abhishek/autotrain-llama3-orpo-v2
|
375 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese
|
376 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
|
@@ -416,6 +440,7 @@
|
|
416 |
- arcee-ai/Arcee-Spark
|
417 |
- arcee-ai/Llama-3.1-SuperNova-Lite
|
418 |
- arcee-ai/SuperNova-Medius
|
|
|
419 |
- arcee-ai/Virtuoso-Small
|
420 |
- argilla/CapybaraHermes-2.5-Mistral-7B
|
421 |
- argilla/notus-7b-v1
|
@@ -443,6 +468,7 @@
|
|
443 |
- botbot-ai/CabraLlama3-8b
|
444 |
- botbot-ai/CabraMistral-v3-7b-32k
|
445 |
- botbot-ai/CabraMixtral-8x7b
|
|
|
446 |
- byroneverson/Mistral-Small-Instruct-2409-abliterated
|
447 |
- byroneverson/Yi-1.5-9B-Chat-16K-abliterated
|
448 |
- byroneverson/Yi-1.5-9B-Chat-abliterated
|
@@ -460,6 +486,10 @@
|
|
460 |
- chujiezheng/Mistral7B-PairRM-SPPO-ExPO
|
461 |
- chujiezheng/Smaug-34B-v0.1-ExPO
|
462 |
- cnmoro/Mistral-7B-Portuguese
|
|
|
|
|
|
|
|
|
463 |
- cognitivecomputations/WestLake-7B-v2-laser
|
464 |
- cognitivecomputations/WizardLM-13B-Uncensored
|
465 |
- cognitivecomputations/WizardLM-30B-Uncensored
|
@@ -481,6 +511,7 @@
|
|
481 |
- cognitivess/bella-1-8b
|
482 |
- cosmicvalor/mistral-orthogonalized
|
483 |
- croissantllm/CroissantLLMBase
|
|
|
484 |
- deepseek-ai/DeepSeek-R1-Distill-Llama-8B
|
485 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
486 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
|
@@ -544,6 +575,7 @@
|
|
544 |
- freewheelin/free-solar-evo-v0.1
|
545 |
- freewheelin/free-solar-evo-v0.11
|
546 |
- freewheelin/free-solar-evo-v0.13
|
|
|
547 |
- ghost-x/ghost-8b-beta
|
548 |
- ghost-x/ghost-8b-beta-1608
|
549 |
- google/gemma-1.1-2b-it
|
@@ -704,6 +736,7 @@
|
|
704 |
- microsoft/phi-1_5
|
705 |
- microsoft/phi-2
|
706 |
- microsoft/phi-4
|
|
|
707 |
- migtissera/Tess-M-v1.3
|
708 |
- migtissera/Tess-v2.5-Gemma-2-27B-alpha
|
709 |
- migtissera/Tess-v2.5-Phi-3-medium-128k-14B
|
@@ -737,12 +770,14 @@
|
|
737 |
- mlabonne/NeuralDaredevil-8B-abliterated
|
738 |
- mlabonne/NeuralMonarch-7B
|
739 |
- monilouise/opt125M_portuguese
|
|
|
740 |
- mosaicml/mpt-30b
|
741 |
- mosaicml/mpt-7b
|
742 |
- mosaicml/mpt-7b-8k
|
743 |
- natong19/Mistral-Nemo-Instruct-2407-abliterated
|
744 |
- natong19/Qwen2-7B-Instruct-abliterated
|
745 |
- nbeerbower/gemma2-gutenberg-27B
|
|
|
746 |
- nicholasKluge/Aira-2-portuguese-124M
|
747 |
- nicholasKluge/Aira-2-portuguese-1B7
|
748 |
- nicholasKluge/Aira-2-portuguese-560M
|
@@ -826,6 +861,7 @@
|
|
826 |
- princeton-nlp/Mistral-7B-Instruct-SimPO
|
827 |
- princeton-nlp/gemma-2-9b-it-DPO
|
828 |
- princeton-nlp/gemma-2-9b-it-SimPO
|
|
|
829 |
- prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
|
830 |
- projecte-aina/FLOR-1.3B
|
831 |
- projecte-aina/FLOR-6.3B
|
@@ -904,11 +940,13 @@
|
|
904 |
- shadowml/BeagSake-7B
|
905 |
- shadowml/Mixolar-4x7b
|
906 |
- sometimesanotion/Lamarck-14B-v0.6
|
|
|
907 |
- sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
|
908 |
- speakleash/Bielik-11B-v2
|
909 |
- speakleash/Bielik-11B-v2.0-Instruct
|
910 |
- speakleash/Bielik-11B-v2.1-Instruct
|
911 |
- speakleash/Bielik-11B-v2.2-Instruct
|
|
|
912 |
- ssmits/Falcon2-5.5B-Portuguese
|
913 |
- ssmits/Falcon2-5.5B-multilingual
|
914 |
- stabilityai/stablelm-2-12b
|
|
|
84 |
- EleutherAI/gpt-neo-125m
|
85 |
- EleutherAI/gpt-neo-2.7B
|
86 |
- EleutherAI/gpt-neox-20b
|
87 |
+
- EleutherAI/polyglot-ko-12.8b
|
88 |
- EleutherAI/pythia-12b
|
89 |
- EleutherAI/pythia-14m
|
90 |
- EleutherAI/pythia-160m
|
|
|
107 |
- EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
|
108 |
- EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
|
109 |
- EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
|
110 |
+
- EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos
|
111 |
- EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
|
112 |
- Eurdem/Defne-llama3.1-8B
|
113 |
+
- FluxiIA/Tucaninho
|
114 |
+
- FluxiIA/Tucaninho_dpo
|
115 |
- FuseAI/FuseChat-7B-VaRM
|
116 |
- FuseAI/OpenChat-3.5-7B-Solar
|
117 |
- GeneZC/MiniChat-1.5-3B
|
|
|
176 |
- Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
|
177 |
- Magpie-Align/MagpieLM-8B-Chat-v0.1
|
178 |
- Magpie-Align/MagpieLM-8B-SFT-v0.1
|
179 |
+
- MagusCorp/grpo_lora_enem_llama3_7b
|
180 |
- MagusCorp/legislinho
|
181 |
- MarinaraSpaghetti/NemoReRemix-12B
|
182 |
- MaziyarPanahi/Calme-4x7B-MoE-v0.1
|
|
|
188 |
- MaziyarPanahi/Mistral-7B-Instruct-v0.3
|
189 |
- MaziyarPanahi/Mistral-7B-v0.3
|
190 |
- MaziyarPanahi/Topxtral-4x7B-v0.1
|
191 |
+
- MaziyarPanahi/calme-2.1-qwen2-7b
|
192 |
+
- MaziyarPanahi/calme-2.2-qwen2-7b
|
193 |
+
- MaziyarPanahi/calme-2.3-qwen2-7b
|
194 |
+
- MaziyarPanahi/calme-2.4-qwen2-7b
|
195 |
- MaziyarPanahi/calme-2.7-qwen2-7b
|
196 |
- MulaBR/Mula-4x160-v0.1
|
197 |
- MulaBR/Mula-8x160-v0.1
|
|
|
201 |
- NTQAI/Nxcode-CQ-7B-orpo
|
202 |
- Nexusflow/Starling-LM-7B-beta
|
203 |
- Nos-PT/Carvalho_pt-gl-1.3B
|
204 |
+
- Nos-PT/Llama-Carvalho-PT
|
205 |
+
- Nos-PT/Llama-Carvalho-PT-GL
|
206 |
- NotAiLOL/Yi-1.5-dolphin-9B
|
207 |
- NousResearch/Hermes-2-Pro-Llama-3-8B
|
208 |
- NousResearch/Hermes-2-Theta-Llama-3-8B
|
|
|
275 |
- Qwen/Qwen2.5-1.5B-Instruct
|
276 |
- Qwen/Qwen2.5-14B
|
277 |
- Qwen/Qwen2.5-14B-Instruct
|
278 |
+
- Qwen/Qwen2.5-14B-Instruct-1M
|
279 |
- Qwen/Qwen2.5-32B
|
280 |
- Qwen/Qwen2.5-32B-Instruct
|
281 |
- Qwen/Qwen2.5-3B
|
|
|
290 |
- Qwen/Qwen2.5-Coder-32B-Instruct
|
291 |
- Qwen/Qwen2.5-Coder-7B
|
292 |
- Qwen/Qwen2.5-Coder-7B-Instruct
|
293 |
+
- Qwen/Qwen2.5-VL-32B-Instruct
|
294 |
+
- Qwen/Qwen2.5-VL-72B-Instruct
|
295 |
+
- RDson/WomboCombo-R1-Coder-14B-Preview
|
296 |
+
- RLHFlow/ArmoRM-Llama3-8B-v0.1
|
297 |
- RLHFlow/LLaMA3-iterative-DPO-final
|
298 |
- Ramikan-BR/Qwen2-0.5B-v5
|
299 |
+
- Ramikan-BR/Qwen2-0.5B-v9
|
300 |
- Ramikan-BR/TiamaPY-v30
|
301 |
- Ramikan-BR/TiamaPY-v31
|
302 |
- Ramikan-BR/tinyllama-coder-py-4bit-v10
|
|
|
308 |
- RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
|
309 |
- Sakalti/Oxyge1-33B
|
310 |
- SakanaAI/DiscoPOP-zephyr-7b-gemma
|
311 |
+
- Salesforce/SFR-Iterative-DPO-LLaMA-3-8B-R
|
312 |
- Sao10K/L3-8B-Niitama-v1
|
313 |
+
- Sao10K/L3.1-8B-Niitama-v1.1
|
314 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
|
315 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
|
316 |
- SeaLLMs/SeaLLM-7B-v2
|
|
|
337 |
- TheBloke/wizardLM-7B-HF
|
338 |
- TheBloke/zephyr-7B-beta-GPTQ
|
339 |
- TheDrummer/Big-Tiger-Gemma-27B-v1
|
340 |
+
- TheDrummer/Gemmasutra-9B-v1
|
341 |
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
|
342 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
343 |
- TinyLlama/TinyLlama_v1.1
|
|
|
359 |
- VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
|
360 |
- VAGOsolutions/SauerkrautLM-Gemma-7b
|
361 |
- VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
|
362 |
+
- VAGOsolutions/SauerkrautLM-Phi-3-medium
|
363 |
- VAGOsolutions/SauerkrautLM-Qwen-32b
|
364 |
- ValiantLabs/Llama3.1-8B-Cobalt
|
365 |
- ValiantLabs/Llama3.1-8B-Fireplace2
|
|
|
389 |
- Xwin-LM/Xwin-LM-7B-V0.2
|
390 |
- abacusai/Liberated-Qwen1.5-14B
|
391 |
- abacusai/Llama-3-Smaug-8B
|
392 |
+
- abacusai/Slerp-CM-mist-dpo
|
393 |
- abacusai/Smaug-34B-v0.1
|
394 |
- abacusai/Smaug-72B-v0.1
|
395 |
- abacusai/Smaug-Llama-3-70B-Instruct
|
396 |
+
- abacusai/bigstral-12b-32k
|
397 |
+
- abacusai/bigyi-15b
|
398 |
- abhishek/autotrain-llama3-orpo-v2
|
399 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese
|
400 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
|
|
|
440 |
- arcee-ai/Arcee-Spark
|
441 |
- arcee-ai/Llama-3.1-SuperNova-Lite
|
442 |
- arcee-ai/SuperNova-Medius
|
443 |
+
- arcee-ai/Virtuoso-Lite
|
444 |
- arcee-ai/Virtuoso-Small
|
445 |
- argilla/CapybaraHermes-2.5-Mistral-7B
|
446 |
- argilla/notus-7b-v1
|
|
|
468 |
- botbot-ai/CabraLlama3-8b
|
469 |
- botbot-ai/CabraMistral-v3-7b-32k
|
470 |
- botbot-ai/CabraMixtral-8x7b
|
471 |
+
- bunnycore/HyperLlama-3.1-8B
|
472 |
- byroneverson/Mistral-Small-Instruct-2409-abliterated
|
473 |
- byroneverson/Yi-1.5-9B-Chat-16K-abliterated
|
474 |
- byroneverson/Yi-1.5-9B-Chat-abliterated
|
|
|
486 |
- chujiezheng/Mistral7B-PairRM-SPPO-ExPO
|
487 |
- chujiezheng/Smaug-34B-v0.1-ExPO
|
488 |
- cnmoro/Mistral-7B-Portuguese
|
489 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-Hybrid-Reasoning
|
490 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-v1
|
491 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-v2
|
492 |
+
- cognitivecomputations/Dolphin3.0-R1-Mistral-24B
|
493 |
- cognitivecomputations/WestLake-7B-v2-laser
|
494 |
- cognitivecomputations/WizardLM-13B-Uncensored
|
495 |
- cognitivecomputations/WizardLM-30B-Uncensored
|
|
|
511 |
- cognitivess/bella-1-8b
|
512 |
- cosmicvalor/mistral-orthogonalized
|
513 |
- croissantllm/CroissantLLMBase
|
514 |
+
- deepseek-ai/DeepSeek-R1-Distill-Llama-70B
|
515 |
- deepseek-ai/DeepSeek-R1-Distill-Llama-8B
|
516 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
517 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
|
|
|
575 |
- freewheelin/free-solar-evo-v0.1
|
576 |
- freewheelin/free-solar-evo-v0.11
|
577 |
- freewheelin/free-solar-evo-v0.13
|
578 |
+
- gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
|
579 |
- ghost-x/ghost-8b-beta
|
580 |
- ghost-x/ghost-8b-beta-1608
|
581 |
- google/gemma-1.1-2b-it
|
|
|
736 |
- microsoft/phi-1_5
|
737 |
- microsoft/phi-2
|
738 |
- microsoft/phi-4
|
739 |
+
- migtissera/Tess-3-Mistral-Nemo-12B
|
740 |
- migtissera/Tess-M-v1.3
|
741 |
- migtissera/Tess-v2.5-Gemma-2-27B-alpha
|
742 |
- migtissera/Tess-v2.5-Phi-3-medium-128k-14B
|
|
|
770 |
- mlabonne/NeuralDaredevil-8B-abliterated
|
771 |
- mlabonne/NeuralMonarch-7B
|
772 |
- monilouise/opt125M_portuguese
|
773 |
+
- monsterapi/Llama-3_1-8B-Instruct-orca-ORPO
|
774 |
- mosaicml/mpt-30b
|
775 |
- mosaicml/mpt-7b
|
776 |
- mosaicml/mpt-7b-8k
|
777 |
- natong19/Mistral-Nemo-Instruct-2407-abliterated
|
778 |
- natong19/Qwen2-7B-Instruct-abliterated
|
779 |
- nbeerbower/gemma2-gutenberg-27B
|
780 |
+
- nbeerbower/mistral-nemo-wissenschaft-12B
|
781 |
- nicholasKluge/Aira-2-portuguese-124M
|
782 |
- nicholasKluge/Aira-2-portuguese-1B7
|
783 |
- nicholasKluge/Aira-2-portuguese-560M
|
|
|
861 |
- princeton-nlp/Mistral-7B-Instruct-SimPO
|
862 |
- princeton-nlp/gemma-2-9b-it-DPO
|
863 |
- princeton-nlp/gemma-2-9b-it-SimPO
|
864 |
+
- prithivMLmods/Megatron-Opus-14B-Exp
|
865 |
- prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
|
866 |
- projecte-aina/FLOR-1.3B
|
867 |
- projecte-aina/FLOR-6.3B
|
|
|
940 |
- shadowml/BeagSake-7B
|
941 |
- shadowml/Mixolar-4x7b
|
942 |
- sometimesanotion/Lamarck-14B-v0.6
|
943 |
+
- sometimesanotion/Lamarck-14B-v0.7
|
944 |
- sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
|
945 |
- speakleash/Bielik-11B-v2
|
946 |
- speakleash/Bielik-11B-v2.0-Instruct
|
947 |
- speakleash/Bielik-11B-v2.1-Instruct
|
948 |
- speakleash/Bielik-11B-v2.2-Instruct
|
949 |
+
- speakleash/Bielik-11B-v2.3-Instruct
|
950 |
- ssmits/Falcon2-5.5B-Portuguese
|
951 |
- ssmits/Falcon2-5.5B-multilingual
|
952 |
- stabilityai/stablelm-2-12b
|
update_models_in_readme.py
CHANGED
@@ -18,8 +18,12 @@ import json
|
|
18 |
|
19 |
snapshot_download(repo_id=QUEUE_REPO, local_dir=EVAL_REQUESTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30)
|
20 |
all_models = []
|
21 |
-
|
22 |
-
|
|
|
|
|
|
|
|
|
23 |
for filepath in glob.glob(os.path.join(EVAL_REQUESTS_PATH, '**/*.json'), recursive=True):
|
24 |
with open(filepath, 'r') as f:
|
25 |
model_data = json.load(f)
|
|
|
18 |
|
19 |
snapshot_download(repo_id=QUEUE_REPO, local_dir=EVAL_REQUESTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30)
|
20 |
all_models = []
|
21 |
+
with open('external_models_results.json', 'r') as f:
|
22 |
+
external_models = json.load(f)
|
23 |
+
for model in external_models:
|
24 |
+
if 'huggingface.co/' in model['link']:
|
25 |
+
all_models.append(model['link'].split('huggingface.co/')[-1])
|
26 |
+
|
27 |
for filepath in glob.glob(os.path.join(EVAL_REQUESTS_PATH, '**/*.json'), recursive=True):
|
28 |
with open(filepath, 'r') as f:
|
29 |
model_data = json.load(f)
|