eduagarcia commited on
Commit
78e9bc6
·
1 Parent(s): b9392b7

New API models

Browse files
README.md CHANGED
@@ -113,6 +113,7 @@ models:
113
  - EleutherAI/gpt-neo-125m
114
  - EleutherAI/gpt-neo-2.7B
115
  - EleutherAI/gpt-neox-20b
 
116
  - EleutherAI/pythia-12b
117
  - EleutherAI/pythia-14m
118
  - EleutherAI/pythia-160m
@@ -135,8 +136,11 @@ models:
135
  - EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
136
  - EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
137
  - EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
 
138
  - EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
139
  - Eurdem/Defne-llama3.1-8B
 
 
140
  - FuseAI/FuseChat-7B-VaRM
141
  - FuseAI/OpenChat-3.5-7B-Solar
142
  - GeneZC/MiniChat-1.5-3B
@@ -201,6 +205,7 @@ models:
201
  - Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
202
  - Magpie-Align/MagpieLM-8B-Chat-v0.1
203
  - Magpie-Align/MagpieLM-8B-SFT-v0.1
 
204
  - MagusCorp/legislinho
205
  - MarinaraSpaghetti/NemoReRemix-12B
206
  - MaziyarPanahi/Calme-4x7B-MoE-v0.1
@@ -212,6 +217,10 @@ models:
212
  - MaziyarPanahi/Mistral-7B-Instruct-v0.3
213
  - MaziyarPanahi/Mistral-7B-v0.3
214
  - MaziyarPanahi/Topxtral-4x7B-v0.1
 
 
 
 
215
  - MaziyarPanahi/calme-2.7-qwen2-7b
216
  - MulaBR/Mula-4x160-v0.1
217
  - MulaBR/Mula-8x160-v0.1
@@ -221,6 +230,8 @@ models:
221
  - NTQAI/Nxcode-CQ-7B-orpo
222
  - Nexusflow/Starling-LM-7B-beta
223
  - Nos-PT/Carvalho_pt-gl-1.3B
 
 
224
  - NotAiLOL/Yi-1.5-dolphin-9B
225
  - NousResearch/Hermes-2-Pro-Llama-3-8B
226
  - NousResearch/Hermes-2-Theta-Llama-3-8B
@@ -293,6 +304,7 @@ models:
293
  - Qwen/Qwen2.5-1.5B-Instruct
294
  - Qwen/Qwen2.5-14B
295
  - Qwen/Qwen2.5-14B-Instruct
 
296
  - Qwen/Qwen2.5-32B
297
  - Qwen/Qwen2.5-32B-Instruct
298
  - Qwen/Qwen2.5-3B
@@ -307,8 +319,13 @@ models:
307
  - Qwen/Qwen2.5-Coder-32B-Instruct
308
  - Qwen/Qwen2.5-Coder-7B
309
  - Qwen/Qwen2.5-Coder-7B-Instruct
 
 
 
 
310
  - RLHFlow/LLaMA3-iterative-DPO-final
311
  - Ramikan-BR/Qwen2-0.5B-v5
 
312
  - Ramikan-BR/TiamaPY-v30
313
  - Ramikan-BR/TiamaPY-v31
314
  - Ramikan-BR/tinyllama-coder-py-4bit-v10
@@ -320,7 +337,9 @@ models:
320
  - RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
321
  - Sakalti/Oxyge1-33B
322
  - SakanaAI/DiscoPOP-zephyr-7b-gemma
 
323
  - Sao10K/L3-8B-Niitama-v1
 
324
  - Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
325
  - Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
326
  - SeaLLMs/SeaLLM-7B-v2
@@ -347,6 +366,7 @@ models:
347
  - TheBloke/wizardLM-7B-HF
348
  - TheBloke/zephyr-7B-beta-GPTQ
349
  - TheDrummer/Big-Tiger-Gemma-27B-v1
 
350
  - TinyLlama/TinyLlama-1.1B-Chat-v1.0
351
  - TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
352
  - TinyLlama/TinyLlama_v1.1
@@ -368,6 +388,7 @@ models:
368
  - VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
369
  - VAGOsolutions/SauerkrautLM-Gemma-7b
370
  - VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
 
371
  - VAGOsolutions/SauerkrautLM-Qwen-32b
372
  - ValiantLabs/Llama3.1-8B-Cobalt
373
  - ValiantLabs/Llama3.1-8B-Fireplace2
@@ -397,9 +418,12 @@ models:
397
  - Xwin-LM/Xwin-LM-7B-V0.2
398
  - abacusai/Liberated-Qwen1.5-14B
399
  - abacusai/Llama-3-Smaug-8B
 
400
  - abacusai/Smaug-34B-v0.1
401
  - abacusai/Smaug-72B-v0.1
402
  - abacusai/Smaug-Llama-3-70B-Instruct
 
 
403
  - abhishek/autotrain-llama3-orpo-v2
404
  - adalbertojunior/Llama-3-8B-Dolphin-Portuguese
405
  - adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
@@ -445,6 +469,7 @@ models:
445
  - arcee-ai/Arcee-Spark
446
  - arcee-ai/Llama-3.1-SuperNova-Lite
447
  - arcee-ai/SuperNova-Medius
 
448
  - arcee-ai/Virtuoso-Small
449
  - argilla/CapybaraHermes-2.5-Mistral-7B
450
  - argilla/notus-7b-v1
@@ -472,6 +497,7 @@ models:
472
  - botbot-ai/CabraLlama3-8b
473
  - botbot-ai/CabraMistral-v3-7b-32k
474
  - botbot-ai/CabraMixtral-8x7b
 
475
  - byroneverson/Mistral-Small-Instruct-2409-abliterated
476
  - byroneverson/Yi-1.5-9B-Chat-16K-abliterated
477
  - byroneverson/Yi-1.5-9B-Chat-abliterated
@@ -489,6 +515,10 @@ models:
489
  - chujiezheng/Mistral7B-PairRM-SPPO-ExPO
490
  - chujiezheng/Smaug-34B-v0.1-ExPO
491
  - cnmoro/Mistral-7B-Portuguese
 
 
 
 
492
  - cognitivecomputations/WestLake-7B-v2-laser
493
  - cognitivecomputations/WizardLM-13B-Uncensored
494
  - cognitivecomputations/WizardLM-30B-Uncensored
@@ -510,6 +540,7 @@ models:
510
  - cognitivess/bella-1-8b
511
  - cosmicvalor/mistral-orthogonalized
512
  - croissantllm/CroissantLLMBase
 
513
  - deepseek-ai/DeepSeek-R1-Distill-Llama-8B
514
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
515
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
@@ -573,6 +604,7 @@ models:
573
  - freewheelin/free-solar-evo-v0.1
574
  - freewheelin/free-solar-evo-v0.11
575
  - freewheelin/free-solar-evo-v0.13
 
576
  - ghost-x/ghost-8b-beta
577
  - ghost-x/ghost-8b-beta-1608
578
  - google/gemma-1.1-2b-it
@@ -733,6 +765,7 @@ models:
733
  - microsoft/phi-1_5
734
  - microsoft/phi-2
735
  - microsoft/phi-4
 
736
  - migtissera/Tess-M-v1.3
737
  - migtissera/Tess-v2.5-Gemma-2-27B-alpha
738
  - migtissera/Tess-v2.5-Phi-3-medium-128k-14B
@@ -766,12 +799,14 @@ models:
766
  - mlabonne/NeuralDaredevil-8B-abliterated
767
  - mlabonne/NeuralMonarch-7B
768
  - monilouise/opt125M_portuguese
 
769
  - mosaicml/mpt-30b
770
  - mosaicml/mpt-7b
771
  - mosaicml/mpt-7b-8k
772
  - natong19/Mistral-Nemo-Instruct-2407-abliterated
773
  - natong19/Qwen2-7B-Instruct-abliterated
774
  - nbeerbower/gemma2-gutenberg-27B
 
775
  - nicholasKluge/Aira-2-portuguese-124M
776
  - nicholasKluge/Aira-2-portuguese-1B7
777
  - nicholasKluge/Aira-2-portuguese-560M
@@ -855,6 +890,7 @@ models:
855
  - princeton-nlp/Mistral-7B-Instruct-SimPO
856
  - princeton-nlp/gemma-2-9b-it-DPO
857
  - princeton-nlp/gemma-2-9b-it-SimPO
 
858
  - prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
859
  - projecte-aina/FLOR-1.3B
860
  - projecte-aina/FLOR-6.3B
@@ -933,11 +969,13 @@ models:
933
  - shadowml/BeagSake-7B
934
  - shadowml/Mixolar-4x7b
935
  - sometimesanotion/Lamarck-14B-v0.6
 
936
  - sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
937
  - speakleash/Bielik-11B-v2
938
  - speakleash/Bielik-11B-v2.0-Instruct
939
  - speakleash/Bielik-11B-v2.1-Instruct
940
  - speakleash/Bielik-11B-v2.2-Instruct
 
941
  - ssmits/Falcon2-5.5B-Portuguese
942
  - ssmits/Falcon2-5.5B-multilingual
943
  - stabilityai/stablelm-2-12b
 
113
  - EleutherAI/gpt-neo-125m
114
  - EleutherAI/gpt-neo-2.7B
115
  - EleutherAI/gpt-neox-20b
116
+ - EleutherAI/polyglot-ko-12.8b
117
  - EleutherAI/pythia-12b
118
  - EleutherAI/pythia-14m
119
  - EleutherAI/pythia-160m
 
136
  - EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
137
  - EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
138
  - EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
139
+ - EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos
140
  - EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
141
  - Eurdem/Defne-llama3.1-8B
142
+ - FluxiIA/Tucaninho
143
+ - FluxiIA/Tucaninho_dpo
144
  - FuseAI/FuseChat-7B-VaRM
145
  - FuseAI/OpenChat-3.5-7B-Solar
146
  - GeneZC/MiniChat-1.5-3B
 
205
  - Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
206
  - Magpie-Align/MagpieLM-8B-Chat-v0.1
207
  - Magpie-Align/MagpieLM-8B-SFT-v0.1
208
+ - MagusCorp/grpo_lora_enem_llama3_7b
209
  - MagusCorp/legislinho
210
  - MarinaraSpaghetti/NemoReRemix-12B
211
  - MaziyarPanahi/Calme-4x7B-MoE-v0.1
 
217
  - MaziyarPanahi/Mistral-7B-Instruct-v0.3
218
  - MaziyarPanahi/Mistral-7B-v0.3
219
  - MaziyarPanahi/Topxtral-4x7B-v0.1
220
+ - MaziyarPanahi/calme-2.1-qwen2-7b
221
+ - MaziyarPanahi/calme-2.2-qwen2-7b
222
+ - MaziyarPanahi/calme-2.3-qwen2-7b
223
+ - MaziyarPanahi/calme-2.4-qwen2-7b
224
  - MaziyarPanahi/calme-2.7-qwen2-7b
225
  - MulaBR/Mula-4x160-v0.1
226
  - MulaBR/Mula-8x160-v0.1
 
230
  - NTQAI/Nxcode-CQ-7B-orpo
231
  - Nexusflow/Starling-LM-7B-beta
232
  - Nos-PT/Carvalho_pt-gl-1.3B
233
+ - Nos-PT/Llama-Carvalho-PT
234
+ - Nos-PT/Llama-Carvalho-PT-GL
235
  - NotAiLOL/Yi-1.5-dolphin-9B
236
  - NousResearch/Hermes-2-Pro-Llama-3-8B
237
  - NousResearch/Hermes-2-Theta-Llama-3-8B
 
304
  - Qwen/Qwen2.5-1.5B-Instruct
305
  - Qwen/Qwen2.5-14B
306
  - Qwen/Qwen2.5-14B-Instruct
307
+ - Qwen/Qwen2.5-14B-Instruct-1M
308
  - Qwen/Qwen2.5-32B
309
  - Qwen/Qwen2.5-32B-Instruct
310
  - Qwen/Qwen2.5-3B
 
319
  - Qwen/Qwen2.5-Coder-32B-Instruct
320
  - Qwen/Qwen2.5-Coder-7B
321
  - Qwen/Qwen2.5-Coder-7B-Instruct
322
+ - Qwen/Qwen2.5-VL-32B-Instruct
323
+ - Qwen/Qwen2.5-VL-72B-Instruct
324
+ - RDson/WomboCombo-R1-Coder-14B-Preview
325
+ - RLHFlow/ArmoRM-Llama3-8B-v0.1
326
  - RLHFlow/LLaMA3-iterative-DPO-final
327
  - Ramikan-BR/Qwen2-0.5B-v5
328
+ - Ramikan-BR/Qwen2-0.5B-v9
329
  - Ramikan-BR/TiamaPY-v30
330
  - Ramikan-BR/TiamaPY-v31
331
  - Ramikan-BR/tinyllama-coder-py-4bit-v10
 
337
  - RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
338
  - Sakalti/Oxyge1-33B
339
  - SakanaAI/DiscoPOP-zephyr-7b-gemma
340
+ - Salesforce/SFR-Iterative-DPO-LLaMA-3-8B-R
341
  - Sao10K/L3-8B-Niitama-v1
342
+ - Sao10K/L3.1-8B-Niitama-v1.1
343
  - Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
344
  - Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
345
  - SeaLLMs/SeaLLM-7B-v2
 
366
  - TheBloke/wizardLM-7B-HF
367
  - TheBloke/zephyr-7B-beta-GPTQ
368
  - TheDrummer/Big-Tiger-Gemma-27B-v1
369
+ - TheDrummer/Gemmasutra-9B-v1
370
  - TinyLlama/TinyLlama-1.1B-Chat-v1.0
371
  - TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
372
  - TinyLlama/TinyLlama_v1.1
 
388
  - VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
389
  - VAGOsolutions/SauerkrautLM-Gemma-7b
390
  - VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
391
+ - VAGOsolutions/SauerkrautLM-Phi-3-medium
392
  - VAGOsolutions/SauerkrautLM-Qwen-32b
393
  - ValiantLabs/Llama3.1-8B-Cobalt
394
  - ValiantLabs/Llama3.1-8B-Fireplace2
 
418
  - Xwin-LM/Xwin-LM-7B-V0.2
419
  - abacusai/Liberated-Qwen1.5-14B
420
  - abacusai/Llama-3-Smaug-8B
421
+ - abacusai/Slerp-CM-mist-dpo
422
  - abacusai/Smaug-34B-v0.1
423
  - abacusai/Smaug-72B-v0.1
424
  - abacusai/Smaug-Llama-3-70B-Instruct
425
+ - abacusai/bigstral-12b-32k
426
+ - abacusai/bigyi-15b
427
  - abhishek/autotrain-llama3-orpo-v2
428
  - adalbertojunior/Llama-3-8B-Dolphin-Portuguese
429
  - adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
 
469
  - arcee-ai/Arcee-Spark
470
  - arcee-ai/Llama-3.1-SuperNova-Lite
471
  - arcee-ai/SuperNova-Medius
472
+ - arcee-ai/Virtuoso-Lite
473
  - arcee-ai/Virtuoso-Small
474
  - argilla/CapybaraHermes-2.5-Mistral-7B
475
  - argilla/notus-7b-v1
 
497
  - botbot-ai/CabraLlama3-8b
498
  - botbot-ai/CabraMistral-v3-7b-32k
499
  - botbot-ai/CabraMixtral-8x7b
500
+ - bunnycore/HyperLlama-3.1-8B
501
  - byroneverson/Mistral-Small-Instruct-2409-abliterated
502
  - byroneverson/Yi-1.5-9B-Chat-16K-abliterated
503
  - byroneverson/Yi-1.5-9B-Chat-abliterated
 
515
  - chujiezheng/Mistral7B-PairRM-SPPO-ExPO
516
  - chujiezheng/Smaug-34B-v0.1-ExPO
517
  - cnmoro/Mistral-7B-Portuguese
518
+ - cnmoro/Qwen2.5-0.5B-Portuguese-Hybrid-Reasoning
519
+ - cnmoro/Qwen2.5-0.5B-Portuguese-v1
520
+ - cnmoro/Qwen2.5-0.5B-Portuguese-v2
521
+ - cognitivecomputations/Dolphin3.0-R1-Mistral-24B
522
  - cognitivecomputations/WestLake-7B-v2-laser
523
  - cognitivecomputations/WizardLM-13B-Uncensored
524
  - cognitivecomputations/WizardLM-30B-Uncensored
 
540
  - cognitivess/bella-1-8b
541
  - cosmicvalor/mistral-orthogonalized
542
  - croissantllm/CroissantLLMBase
543
+ - deepseek-ai/DeepSeek-R1-Distill-Llama-70B
544
  - deepseek-ai/DeepSeek-R1-Distill-Llama-8B
545
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
546
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
 
604
  - freewheelin/free-solar-evo-v0.1
605
  - freewheelin/free-solar-evo-v0.11
606
  - freewheelin/free-solar-evo-v0.13
607
+ - gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
608
  - ghost-x/ghost-8b-beta
609
  - ghost-x/ghost-8b-beta-1608
610
  - google/gemma-1.1-2b-it
 
765
  - microsoft/phi-1_5
766
  - microsoft/phi-2
767
  - microsoft/phi-4
768
+ - migtissera/Tess-3-Mistral-Nemo-12B
769
  - migtissera/Tess-M-v1.3
770
  - migtissera/Tess-v2.5-Gemma-2-27B-alpha
771
  - migtissera/Tess-v2.5-Phi-3-medium-128k-14B
 
799
  - mlabonne/NeuralDaredevil-8B-abliterated
800
  - mlabonne/NeuralMonarch-7B
801
  - monilouise/opt125M_portuguese
802
+ - monsterapi/Llama-3_1-8B-Instruct-orca-ORPO
803
  - mosaicml/mpt-30b
804
  - mosaicml/mpt-7b
805
  - mosaicml/mpt-7b-8k
806
  - natong19/Mistral-Nemo-Instruct-2407-abliterated
807
  - natong19/Qwen2-7B-Instruct-abliterated
808
  - nbeerbower/gemma2-gutenberg-27B
809
+ - nbeerbower/mistral-nemo-wissenschaft-12B
810
  - nicholasKluge/Aira-2-portuguese-124M
811
  - nicholasKluge/Aira-2-portuguese-1B7
812
  - nicholasKluge/Aira-2-portuguese-560M
 
890
  - princeton-nlp/Mistral-7B-Instruct-SimPO
891
  - princeton-nlp/gemma-2-9b-it-DPO
892
  - princeton-nlp/gemma-2-9b-it-SimPO
893
+ - prithivMLmods/Megatron-Opus-14B-Exp
894
  - prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
895
  - projecte-aina/FLOR-1.3B
896
  - projecte-aina/FLOR-6.3B
 
969
  - shadowml/BeagSake-7B
970
  - shadowml/Mixolar-4x7b
971
  - sometimesanotion/Lamarck-14B-v0.6
972
+ - sometimesanotion/Lamarck-14B-v0.7
973
  - sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
974
  - speakleash/Bielik-11B-v2
975
  - speakleash/Bielik-11B-v2.0-Instruct
976
  - speakleash/Bielik-11B-v2.1-Instruct
977
  - speakleash/Bielik-11B-v2.2-Instruct
978
+ - speakleash/Bielik-11B-v2.3-Instruct
979
  - ssmits/Falcon2-5.5B-Portuguese
980
  - ssmits/Falcon2-5.5B-multilingual
981
  - stabilityai/stablelm-2-12b
external_models_results.json CHANGED
@@ -176,8 +176,8 @@
176
  "result_metrics_npm": 0.6834036936130392
177
  },
178
  {
179
- "model": "gemini-1.5-flash",
180
- "name": "Gemini 1.5 Flash",
181
  "link": "https://cloud.google.com/vertex-ai",
182
  "date": "2024-08-09",
183
  "status": "full",
@@ -222,7 +222,7 @@
222
  {
223
  "model": "nemotron-4-340b-instruct",
224
  "name": "nvidia/Nemotron-4-340B-Instruct (Nvidia API)",
225
- "link": "https://build.nvidia.com/nvidia/nemotron-4-340b-instruct",
226
  "date": "2024-06-30",
227
  "status": "full",
228
  "main_language": "English",
@@ -333,6 +333,51 @@
333
  "result_metrics_average": 0.7928134532766066,
334
  "result_metrics_npm": 0.6915070359785283
335
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
336
  {
337
  "model": "gemini-2.0-flash-001",
338
  "name": "Gemini 2.0 Flash (001)",
@@ -376,5 +421,115 @@
376
  },
377
  "result_metrics_average": 0.8056048352614735,
378
  "result_metrics_npm": 0.6986042497176748
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
379
  }
380
  ]
 
176
  "result_metrics_npm": 0.6834036936130392
177
  },
178
  {
179
+ "model": "gemini-1.5-flash-001",
180
+ "name": "Gemini 1.5 Flash (001)",
181
  "link": "https://cloud.google.com/vertex-ai",
182
  "date": "2024-08-09",
183
  "status": "full",
 
222
  {
223
  "model": "nemotron-4-340b-instruct",
224
  "name": "nvidia/Nemotron-4-340B-Instruct (Nvidia API)",
225
+ "link": "https://huggingface.co/nvidia/Nemotron-4-340B-Instruct",
226
  "date": "2024-06-30",
227
  "status": "full",
228
  "main_language": "English",
 
333
  "result_metrics_average": 0.7928134532766066,
334
  "result_metrics_npm": 0.6915070359785283
335
  },
336
+ {
337
+ "model": "gemini-1.5-flash-002",
338
+ "name": "Gemini 1.5 Flash (002)",
339
+ "link": "https://cloud.google.com/vertex-ai",
340
+ "date": "2025-04-03",
341
+ "status": "full",
342
+ "main_language": "English",
343
+ "model_type": "proprietary",
344
+ "result_metrics": {
345
+ "enem_challenge": 0.8327501749475158,
346
+ "bluex": 0.760778859527121,
347
+ "oab_exams": 0.6369020501138952,
348
+ "assin2_sts": 0.8380176734291938,
349
+ "assin2_rte": 0.941176117215237,
350
+ "faquad_nli": 0.8360786822325283,
351
+ "hatebr_offensive": 0.9046145161133335,
352
+ "portuguese_hate_speech": 0.7406414313684444,
353
+ "tweetsentbr": 0.6997509880131249
354
+ },
355
+ "result_metrics_average": 0.7989678325511549,
356
+ "result_metrics_npm": 0.6979777100000177
357
+ },
358
+ {
359
+ "model": "gemini-1.5-flash-8b-001",
360
+ "name": "Gemini 1.5 Flash 8B (001)",
361
+ "link": "https://aistudio.google.com",
362
+ "date": "2025-04-03",
363
+ "status": "full",
364
+ "main_language": "English",
365
+ "model_type": "proprietary",
366
+ "params": 8.0,
367
+ "result_metrics": {
368
+ "enem_challenge": 0.7641707487753674,
369
+ "bluex": 0.6467315716272601,
370
+ "oab_exams": 0.5603644646924829,
371
+ "assin2_sts": 0.7638946799836569,
372
+ "assin2_rte": 0.9329452628161146,
373
+ "faquad_nli": 0.7937022965448601,
374
+ "hatebr_offensive": 0.850497640901663,
375
+ "portuguese_hate_speech": 0.7391317606010173,
376
+ "tweetsentbr": 0.7376684798923661
377
+ },
378
+ "result_metrics_average": 0.7543452117594209,
379
+ "result_metrics_npm": 0.6359642422837162
380
+ },
381
  {
382
  "model": "gemini-2.0-flash-001",
383
  "name": "Gemini 2.0 Flash (001)",
 
421
  },
422
  "result_metrics_average": 0.8056048352614735,
423
  "result_metrics_npm": 0.6986042497176748
424
+ },
425
+ {
426
+ "model": "gemini-2.5-pro-exp-03-25",
427
+ "name": "Gemini 2.5 Pro Experimental (0325)",
428
+ "link": "https://aistudio.google.com",
429
+ "date": "2025-04-03",
430
+ "status": "full",
431
+ "main_language": "English",
432
+ "model_type": "proprietary",
433
+ "result_metrics": {
434
+ "enem_challenge": 0.9769069279216235,
435
+ "bluex": 0.9499304589707928,
436
+ "oab_exams": 0.9216400911161731,
437
+ "assin2_sts": 0.837785744915033,
438
+ "assin2_rte": 0.9415510158830285,
439
+ "faquad_nli": 0.8738735797309651,
440
+ "hatebr_offensive": 0.9248478168290788,
441
+ "portuguese_hate_speech": 0.7336133105156697,
442
+ "tweetsentbr": 0.7928002469993594
443
+ },
444
+ "result_metrics_average": 0.8836610214313025,
445
+ "result_metrics_npm": 0.8134610556797854
446
+ },
447
+ {
448
+ "model": "qwen2-5-vl-72b-instruct",
449
+ "name": "Qwen/Qwen2.5-VL-72B-Instruct (API)",
450
+ "link": "https://huggingface.co/Qwen/Qwen2.5-VL-72B-Instruct",
451
+ "date": "2025-04-03",
452
+ "status": "full",
453
+ "main_language": "English",
454
+ "model_type": "chat",
455
+ "result_metrics": {
456
+ "enem_challenge": 0.8600419874037789,
457
+ "bluex": 0.8052851182197497,
458
+ "oab_exams": 0.6888382687927107,
459
+ "assin2_sts": 0.7595538567467497,
460
+ "assin2_rte": 0.9472975104201871,
461
+ "faquad_nli": 0.8447190882122586,
462
+ "hatebr_offensive": 0.8810695094657859,
463
+ "portuguese_hate_speech": 0.769596419318135,
464
+ "tweetsentbr": 0.5644757075411895
465
+ },
466
+ "result_metrics_average": 0.7912086073467273,
467
+ "result_metrics_npm": 0.6888261361422966
468
+ },
469
+ {
470
+ "model": "qwen2-5-72b-instruct",
471
+ "name": "Qwen/Qwen2.5-72B-Instruct (API)",
472
+ "link": "https://huggingface.co/Qwen/Qwen2.5-72B-Instruct",
473
+ "date": "2025-04-03",
474
+ "status": "full",
475
+ "main_language": "English",
476
+ "model_type": "chat",
477
+ "result_metrics": {
478
+ "enem_challenge": 0.8432470258922323,
479
+ "bluex": 0.780250347705146,
480
+ "oab_exams": 0.675626423690205,
481
+ "assin2_sts": 0.8230708844558656,
482
+ "assin2_rte": 0.9509720145268106,
483
+ "faquad_nli": 0.8194444444444444,
484
+ "hatebr_offensive": 0.8810033427242816,
485
+ "portuguese_hate_speech": 0.7601866578782712,
486
+ "tweetsentbr": 0.7620172222071487
487
+ },
488
+ "result_metrics_average": 0.8106464848360451,
489
+ "result_metrics_npm": 0.7142994872542282
490
+ },
491
+ {
492
+ "model": "qwen2-5-vl-32b-instruct",
493
+ "name": "Qwen/Qwen2.5-VL-32B-Instruct (API)",
494
+ "link": "https://huggingface.co/Qwen/Qwen2.5-VL-32B-Instruct",
495
+ "date": "2025-04-03",
496
+ "status": "full",
497
+ "main_language": "English",
498
+ "model_type": "chat",
499
+ "result_metrics": {
500
+ "enem_challenge": 0.8600419874037789,
501
+ "bluex": 0.8052851182197497,
502
+ "oab_exams": 0.6888382687927107,
503
+ "assin2_sts": 0.7780549055529008,
504
+ "assin2_rte": 0.9472975104201871,
505
+ "faquad_nli": 0.8447190882122586,
506
+ "hatebr_offensive": 0.8810695094657859,
507
+ "portuguese_hate_speech": 0.769596419318135,
508
+ "tweetsentbr": 0.7027408707999051
509
+ },
510
+ "result_metrics_average": 0.8086270753539346,
511
+ "result_metrics_npm": 0.7137431116807307
512
+ },
513
+ {
514
+ "model": "qwen-turbo-2024-11-01",
515
+ "name": "Qwen-Turbo (2024-11-01)",
516
+ "link": "https://www.alibabacloud.com/en/product/modelstudio",
517
+ "date": "2025-04-03",
518
+ "status": "full",
519
+ "main_language": "English",
520
+ "model_type": "proprietary",
521
+ "result_metrics": {
522
+ "enem_challenge": 0.7795661301609517,
523
+ "bluex": 0.7079276773296245,
524
+ "oab_exams": 0.6091116173120729,
525
+ "assin2_sts": 0.7640477700456898,
526
+ "assin2_rte": 0.9260451969385788,
527
+ "faquad_nli": 0.8128063725490196,
528
+ "hatebr_offensive": 0.8567933277676292,
529
+ "portuguese_hate_speech": 0.7239183383094245,
530
+ "tweetsentbr": 0.7038360447972195
531
+ },
532
+ "result_metrics_average": 0.7648947194678011,
533
+ "result_metrics_npm": 0.6490441260447987
534
  }
535
  ]
model_list.txt CHANGED
@@ -84,6 +84,7 @@
84
  - EleutherAI/gpt-neo-125m
85
  - EleutherAI/gpt-neo-2.7B
86
  - EleutherAI/gpt-neox-20b
 
87
  - EleutherAI/pythia-12b
88
  - EleutherAI/pythia-14m
89
  - EleutherAI/pythia-160m
@@ -106,8 +107,11 @@
106
  - EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
107
  - EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
108
  - EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
 
109
  - EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
110
  - Eurdem/Defne-llama3.1-8B
 
 
111
  - FuseAI/FuseChat-7B-VaRM
112
  - FuseAI/OpenChat-3.5-7B-Solar
113
  - GeneZC/MiniChat-1.5-3B
@@ -172,6 +176,7 @@
172
  - Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
173
  - Magpie-Align/MagpieLM-8B-Chat-v0.1
174
  - Magpie-Align/MagpieLM-8B-SFT-v0.1
 
175
  - MagusCorp/legislinho
176
  - MarinaraSpaghetti/NemoReRemix-12B
177
  - MaziyarPanahi/Calme-4x7B-MoE-v0.1
@@ -183,6 +188,10 @@
183
  - MaziyarPanahi/Mistral-7B-Instruct-v0.3
184
  - MaziyarPanahi/Mistral-7B-v0.3
185
  - MaziyarPanahi/Topxtral-4x7B-v0.1
 
 
 
 
186
  - MaziyarPanahi/calme-2.7-qwen2-7b
187
  - MulaBR/Mula-4x160-v0.1
188
  - MulaBR/Mula-8x160-v0.1
@@ -192,6 +201,8 @@
192
  - NTQAI/Nxcode-CQ-7B-orpo
193
  - Nexusflow/Starling-LM-7B-beta
194
  - Nos-PT/Carvalho_pt-gl-1.3B
 
 
195
  - NotAiLOL/Yi-1.5-dolphin-9B
196
  - NousResearch/Hermes-2-Pro-Llama-3-8B
197
  - NousResearch/Hermes-2-Theta-Llama-3-8B
@@ -264,6 +275,7 @@
264
  - Qwen/Qwen2.5-1.5B-Instruct
265
  - Qwen/Qwen2.5-14B
266
  - Qwen/Qwen2.5-14B-Instruct
 
267
  - Qwen/Qwen2.5-32B
268
  - Qwen/Qwen2.5-32B-Instruct
269
  - Qwen/Qwen2.5-3B
@@ -278,8 +290,13 @@
278
  - Qwen/Qwen2.5-Coder-32B-Instruct
279
  - Qwen/Qwen2.5-Coder-7B
280
  - Qwen/Qwen2.5-Coder-7B-Instruct
 
 
 
 
281
  - RLHFlow/LLaMA3-iterative-DPO-final
282
  - Ramikan-BR/Qwen2-0.5B-v5
 
283
  - Ramikan-BR/TiamaPY-v30
284
  - Ramikan-BR/TiamaPY-v31
285
  - Ramikan-BR/tinyllama-coder-py-4bit-v10
@@ -291,7 +308,9 @@
291
  - RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
292
  - Sakalti/Oxyge1-33B
293
  - SakanaAI/DiscoPOP-zephyr-7b-gemma
 
294
  - Sao10K/L3-8B-Niitama-v1
 
295
  - Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
296
  - Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
297
  - SeaLLMs/SeaLLM-7B-v2
@@ -318,6 +337,7 @@
318
  - TheBloke/wizardLM-7B-HF
319
  - TheBloke/zephyr-7B-beta-GPTQ
320
  - TheDrummer/Big-Tiger-Gemma-27B-v1
 
321
  - TinyLlama/TinyLlama-1.1B-Chat-v1.0
322
  - TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
323
  - TinyLlama/TinyLlama_v1.1
@@ -339,6 +359,7 @@
339
  - VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
340
  - VAGOsolutions/SauerkrautLM-Gemma-7b
341
  - VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
 
342
  - VAGOsolutions/SauerkrautLM-Qwen-32b
343
  - ValiantLabs/Llama3.1-8B-Cobalt
344
  - ValiantLabs/Llama3.1-8B-Fireplace2
@@ -368,9 +389,12 @@
368
  - Xwin-LM/Xwin-LM-7B-V0.2
369
  - abacusai/Liberated-Qwen1.5-14B
370
  - abacusai/Llama-3-Smaug-8B
 
371
  - abacusai/Smaug-34B-v0.1
372
  - abacusai/Smaug-72B-v0.1
373
  - abacusai/Smaug-Llama-3-70B-Instruct
 
 
374
  - abhishek/autotrain-llama3-orpo-v2
375
  - adalbertojunior/Llama-3-8B-Dolphin-Portuguese
376
  - adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
@@ -416,6 +440,7 @@
416
  - arcee-ai/Arcee-Spark
417
  - arcee-ai/Llama-3.1-SuperNova-Lite
418
  - arcee-ai/SuperNova-Medius
 
419
  - arcee-ai/Virtuoso-Small
420
  - argilla/CapybaraHermes-2.5-Mistral-7B
421
  - argilla/notus-7b-v1
@@ -443,6 +468,7 @@
443
  - botbot-ai/CabraLlama3-8b
444
  - botbot-ai/CabraMistral-v3-7b-32k
445
  - botbot-ai/CabraMixtral-8x7b
 
446
  - byroneverson/Mistral-Small-Instruct-2409-abliterated
447
  - byroneverson/Yi-1.5-9B-Chat-16K-abliterated
448
  - byroneverson/Yi-1.5-9B-Chat-abliterated
@@ -460,6 +486,10 @@
460
  - chujiezheng/Mistral7B-PairRM-SPPO-ExPO
461
  - chujiezheng/Smaug-34B-v0.1-ExPO
462
  - cnmoro/Mistral-7B-Portuguese
 
 
 
 
463
  - cognitivecomputations/WestLake-7B-v2-laser
464
  - cognitivecomputations/WizardLM-13B-Uncensored
465
  - cognitivecomputations/WizardLM-30B-Uncensored
@@ -481,6 +511,7 @@
481
  - cognitivess/bella-1-8b
482
  - cosmicvalor/mistral-orthogonalized
483
  - croissantllm/CroissantLLMBase
 
484
  - deepseek-ai/DeepSeek-R1-Distill-Llama-8B
485
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
486
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
@@ -544,6 +575,7 @@
544
  - freewheelin/free-solar-evo-v0.1
545
  - freewheelin/free-solar-evo-v0.11
546
  - freewheelin/free-solar-evo-v0.13
 
547
  - ghost-x/ghost-8b-beta
548
  - ghost-x/ghost-8b-beta-1608
549
  - google/gemma-1.1-2b-it
@@ -704,6 +736,7 @@
704
  - microsoft/phi-1_5
705
  - microsoft/phi-2
706
  - microsoft/phi-4
 
707
  - migtissera/Tess-M-v1.3
708
  - migtissera/Tess-v2.5-Gemma-2-27B-alpha
709
  - migtissera/Tess-v2.5-Phi-3-medium-128k-14B
@@ -737,12 +770,14 @@
737
  - mlabonne/NeuralDaredevil-8B-abliterated
738
  - mlabonne/NeuralMonarch-7B
739
  - monilouise/opt125M_portuguese
 
740
  - mosaicml/mpt-30b
741
  - mosaicml/mpt-7b
742
  - mosaicml/mpt-7b-8k
743
  - natong19/Mistral-Nemo-Instruct-2407-abliterated
744
  - natong19/Qwen2-7B-Instruct-abliterated
745
  - nbeerbower/gemma2-gutenberg-27B
 
746
  - nicholasKluge/Aira-2-portuguese-124M
747
  - nicholasKluge/Aira-2-portuguese-1B7
748
  - nicholasKluge/Aira-2-portuguese-560M
@@ -826,6 +861,7 @@
826
  - princeton-nlp/Mistral-7B-Instruct-SimPO
827
  - princeton-nlp/gemma-2-9b-it-DPO
828
  - princeton-nlp/gemma-2-9b-it-SimPO
 
829
  - prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
830
  - projecte-aina/FLOR-1.3B
831
  - projecte-aina/FLOR-6.3B
@@ -904,11 +940,13 @@
904
  - shadowml/BeagSake-7B
905
  - shadowml/Mixolar-4x7b
906
  - sometimesanotion/Lamarck-14B-v0.6
 
907
  - sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
908
  - speakleash/Bielik-11B-v2
909
  - speakleash/Bielik-11B-v2.0-Instruct
910
  - speakleash/Bielik-11B-v2.1-Instruct
911
  - speakleash/Bielik-11B-v2.2-Instruct
 
912
  - ssmits/Falcon2-5.5B-Portuguese
913
  - ssmits/Falcon2-5.5B-multilingual
914
  - stabilityai/stablelm-2-12b
 
84
  - EleutherAI/gpt-neo-125m
85
  - EleutherAI/gpt-neo-2.7B
86
  - EleutherAI/gpt-neox-20b
87
+ - EleutherAI/polyglot-ko-12.8b
88
  - EleutherAI/pythia-12b
89
  - EleutherAI/pythia-14m
90
  - EleutherAI/pythia-160m
 
107
  - EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
108
  - EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
109
  - EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
110
+ - EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos
111
  - EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
112
  - Eurdem/Defne-llama3.1-8B
113
+ - FluxiIA/Tucaninho
114
+ - FluxiIA/Tucaninho_dpo
115
  - FuseAI/FuseChat-7B-VaRM
116
  - FuseAI/OpenChat-3.5-7B-Solar
117
  - GeneZC/MiniChat-1.5-3B
 
176
  - Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
177
  - Magpie-Align/MagpieLM-8B-Chat-v0.1
178
  - Magpie-Align/MagpieLM-8B-SFT-v0.1
179
+ - MagusCorp/grpo_lora_enem_llama3_7b
180
  - MagusCorp/legislinho
181
  - MarinaraSpaghetti/NemoReRemix-12B
182
  - MaziyarPanahi/Calme-4x7B-MoE-v0.1
 
188
  - MaziyarPanahi/Mistral-7B-Instruct-v0.3
189
  - MaziyarPanahi/Mistral-7B-v0.3
190
  - MaziyarPanahi/Topxtral-4x7B-v0.1
191
+ - MaziyarPanahi/calme-2.1-qwen2-7b
192
+ - MaziyarPanahi/calme-2.2-qwen2-7b
193
+ - MaziyarPanahi/calme-2.3-qwen2-7b
194
+ - MaziyarPanahi/calme-2.4-qwen2-7b
195
  - MaziyarPanahi/calme-2.7-qwen2-7b
196
  - MulaBR/Mula-4x160-v0.1
197
  - MulaBR/Mula-8x160-v0.1
 
201
  - NTQAI/Nxcode-CQ-7B-orpo
202
  - Nexusflow/Starling-LM-7B-beta
203
  - Nos-PT/Carvalho_pt-gl-1.3B
204
+ - Nos-PT/Llama-Carvalho-PT
205
+ - Nos-PT/Llama-Carvalho-PT-GL
206
  - NotAiLOL/Yi-1.5-dolphin-9B
207
  - NousResearch/Hermes-2-Pro-Llama-3-8B
208
  - NousResearch/Hermes-2-Theta-Llama-3-8B
 
275
  - Qwen/Qwen2.5-1.5B-Instruct
276
  - Qwen/Qwen2.5-14B
277
  - Qwen/Qwen2.5-14B-Instruct
278
+ - Qwen/Qwen2.5-14B-Instruct-1M
279
  - Qwen/Qwen2.5-32B
280
  - Qwen/Qwen2.5-32B-Instruct
281
  - Qwen/Qwen2.5-3B
 
290
  - Qwen/Qwen2.5-Coder-32B-Instruct
291
  - Qwen/Qwen2.5-Coder-7B
292
  - Qwen/Qwen2.5-Coder-7B-Instruct
293
+ - Qwen/Qwen2.5-VL-32B-Instruct
294
+ - Qwen/Qwen2.5-VL-72B-Instruct
295
+ - RDson/WomboCombo-R1-Coder-14B-Preview
296
+ - RLHFlow/ArmoRM-Llama3-8B-v0.1
297
  - RLHFlow/LLaMA3-iterative-DPO-final
298
  - Ramikan-BR/Qwen2-0.5B-v5
299
+ - Ramikan-BR/Qwen2-0.5B-v9
300
  - Ramikan-BR/TiamaPY-v30
301
  - Ramikan-BR/TiamaPY-v31
302
  - Ramikan-BR/tinyllama-coder-py-4bit-v10
 
308
  - RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
309
  - Sakalti/Oxyge1-33B
310
  - SakanaAI/DiscoPOP-zephyr-7b-gemma
311
+ - Salesforce/SFR-Iterative-DPO-LLaMA-3-8B-R
312
  - Sao10K/L3-8B-Niitama-v1
313
+ - Sao10K/L3.1-8B-Niitama-v1.1
314
  - Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
315
  - Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
316
  - SeaLLMs/SeaLLM-7B-v2
 
337
  - TheBloke/wizardLM-7B-HF
338
  - TheBloke/zephyr-7B-beta-GPTQ
339
  - TheDrummer/Big-Tiger-Gemma-27B-v1
340
+ - TheDrummer/Gemmasutra-9B-v1
341
  - TinyLlama/TinyLlama-1.1B-Chat-v1.0
342
  - TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
343
  - TinyLlama/TinyLlama_v1.1
 
359
  - VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
360
  - VAGOsolutions/SauerkrautLM-Gemma-7b
361
  - VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
362
+ - VAGOsolutions/SauerkrautLM-Phi-3-medium
363
  - VAGOsolutions/SauerkrautLM-Qwen-32b
364
  - ValiantLabs/Llama3.1-8B-Cobalt
365
  - ValiantLabs/Llama3.1-8B-Fireplace2
 
389
  - Xwin-LM/Xwin-LM-7B-V0.2
390
  - abacusai/Liberated-Qwen1.5-14B
391
  - abacusai/Llama-3-Smaug-8B
392
+ - abacusai/Slerp-CM-mist-dpo
393
  - abacusai/Smaug-34B-v0.1
394
  - abacusai/Smaug-72B-v0.1
395
  - abacusai/Smaug-Llama-3-70B-Instruct
396
+ - abacusai/bigstral-12b-32k
397
+ - abacusai/bigyi-15b
398
  - abhishek/autotrain-llama3-orpo-v2
399
  - adalbertojunior/Llama-3-8B-Dolphin-Portuguese
400
  - adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
 
440
  - arcee-ai/Arcee-Spark
441
  - arcee-ai/Llama-3.1-SuperNova-Lite
442
  - arcee-ai/SuperNova-Medius
443
+ - arcee-ai/Virtuoso-Lite
444
  - arcee-ai/Virtuoso-Small
445
  - argilla/CapybaraHermes-2.5-Mistral-7B
446
  - argilla/notus-7b-v1
 
468
  - botbot-ai/CabraLlama3-8b
469
  - botbot-ai/CabraMistral-v3-7b-32k
470
  - botbot-ai/CabraMixtral-8x7b
471
+ - bunnycore/HyperLlama-3.1-8B
472
  - byroneverson/Mistral-Small-Instruct-2409-abliterated
473
  - byroneverson/Yi-1.5-9B-Chat-16K-abliterated
474
  - byroneverson/Yi-1.5-9B-Chat-abliterated
 
486
  - chujiezheng/Mistral7B-PairRM-SPPO-ExPO
487
  - chujiezheng/Smaug-34B-v0.1-ExPO
488
  - cnmoro/Mistral-7B-Portuguese
489
+ - cnmoro/Qwen2.5-0.5B-Portuguese-Hybrid-Reasoning
490
+ - cnmoro/Qwen2.5-0.5B-Portuguese-v1
491
+ - cnmoro/Qwen2.5-0.5B-Portuguese-v2
492
+ - cognitivecomputations/Dolphin3.0-R1-Mistral-24B
493
  - cognitivecomputations/WestLake-7B-v2-laser
494
  - cognitivecomputations/WizardLM-13B-Uncensored
495
  - cognitivecomputations/WizardLM-30B-Uncensored
 
511
  - cognitivess/bella-1-8b
512
  - cosmicvalor/mistral-orthogonalized
513
  - croissantllm/CroissantLLMBase
514
+ - deepseek-ai/DeepSeek-R1-Distill-Llama-70B
515
  - deepseek-ai/DeepSeek-R1-Distill-Llama-8B
516
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
517
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
 
575
  - freewheelin/free-solar-evo-v0.1
576
  - freewheelin/free-solar-evo-v0.11
577
  - freewheelin/free-solar-evo-v0.13
578
+ - gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
579
  - ghost-x/ghost-8b-beta
580
  - ghost-x/ghost-8b-beta-1608
581
  - google/gemma-1.1-2b-it
 
736
  - microsoft/phi-1_5
737
  - microsoft/phi-2
738
  - microsoft/phi-4
739
+ - migtissera/Tess-3-Mistral-Nemo-12B
740
  - migtissera/Tess-M-v1.3
741
  - migtissera/Tess-v2.5-Gemma-2-27B-alpha
742
  - migtissera/Tess-v2.5-Phi-3-medium-128k-14B
 
770
  - mlabonne/NeuralDaredevil-8B-abliterated
771
  - mlabonne/NeuralMonarch-7B
772
  - monilouise/opt125M_portuguese
773
+ - monsterapi/Llama-3_1-8B-Instruct-orca-ORPO
774
  - mosaicml/mpt-30b
775
  - mosaicml/mpt-7b
776
  - mosaicml/mpt-7b-8k
777
  - natong19/Mistral-Nemo-Instruct-2407-abliterated
778
  - natong19/Qwen2-7B-Instruct-abliterated
779
  - nbeerbower/gemma2-gutenberg-27B
780
+ - nbeerbower/mistral-nemo-wissenschaft-12B
781
  - nicholasKluge/Aira-2-portuguese-124M
782
  - nicholasKluge/Aira-2-portuguese-1B7
783
  - nicholasKluge/Aira-2-portuguese-560M
 
861
  - princeton-nlp/Mistral-7B-Instruct-SimPO
862
  - princeton-nlp/gemma-2-9b-it-DPO
863
  - princeton-nlp/gemma-2-9b-it-SimPO
864
+ - prithivMLmods/Megatron-Opus-14B-Exp
865
  - prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
866
  - projecte-aina/FLOR-1.3B
867
  - projecte-aina/FLOR-6.3B
 
940
  - shadowml/BeagSake-7B
941
  - shadowml/Mixolar-4x7b
942
  - sometimesanotion/Lamarck-14B-v0.6
943
+ - sometimesanotion/Lamarck-14B-v0.7
944
  - sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
945
  - speakleash/Bielik-11B-v2
946
  - speakleash/Bielik-11B-v2.0-Instruct
947
  - speakleash/Bielik-11B-v2.1-Instruct
948
  - speakleash/Bielik-11B-v2.2-Instruct
949
+ - speakleash/Bielik-11B-v2.3-Instruct
950
  - ssmits/Falcon2-5.5B-Portuguese
951
  - ssmits/Falcon2-5.5B-multilingual
952
  - stabilityai/stablelm-2-12b
update_models_in_readme.py CHANGED
@@ -18,8 +18,12 @@ import json
18
 
19
  snapshot_download(repo_id=QUEUE_REPO, local_dir=EVAL_REQUESTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30)
20
  all_models = []
21
- all_models.append('nvidia/Nemotron-4-340B-Instruct')
22
- all_models.append('meta-llama/Meta-Llama-3.1-405B-Instruct')
 
 
 
 
23
  for filepath in glob.glob(os.path.join(EVAL_REQUESTS_PATH, '**/*.json'), recursive=True):
24
  with open(filepath, 'r') as f:
25
  model_data = json.load(f)
 
18
 
19
  snapshot_download(repo_id=QUEUE_REPO, local_dir=EVAL_REQUESTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30)
20
  all_models = []
21
+ with open('external_models_results.json', 'r') as f:
22
+ external_models = json.load(f)
23
+ for model in external_models:
24
+ if 'huggingface.co/' in model['link']:
25
+ all_models.append(model['link'].split('huggingface.co/')[-1])
26
+
27
  for filepath in glob.glob(os.path.join(EVAL_REQUESTS_PATH, '**/*.json'), recursive=True):
28
  with open(filepath, 'r') as f:
29
  model_data = json.load(f)