CultriX commited on
Commit
6b68288
·
verified ·
1 Parent(s): 6acb805

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +713 -648
app.py CHANGED
@@ -319,520 +319,584 @@ def download_all_data():
319
  # --------------------------------------------------------------------
320
  # This is your larger dataset, rank = 44..105
321
  benchmark_data = [
322
- {
323
- "Model Rank": "1",
324
- "Icon": "ð¤",
325
- "Model Name": "wanlige/li-14b-v0.4",
326
- "Average Score": "43.66 %",
327
- "IFEval Score": "81.33 %",
328
- "BBH Score": "50.38 %",
329
- "MATH Score": "55.74 %",
330
- "GPQA Score": "11.86 %",
331
- "MUSR Score": "16.35 %",
332
- "MMLU-PRO Score": "46.30 %",
333
- "Model Architecture": "Qwen2ForCausalLM",
334
- "Model Precision": "bfloat16",
335
- "Model Parameters": "14.77B",
336
- "Chat Template Use": "Yes"
337
- },
338
- {
339
- "Model Rank": "2",
340
- "Icon": "ð¤",
341
- "Model Name": "suayptalha/Lamarckvergence-14B",
342
- "Average Score": "43.32 %",
343
- "IFEval Score": "76.56 %",
344
- "BBH Score": "50.33 %",
345
- "MATH Score": "54.00 %",
346
- "GPQA Score": "15.10 %",
347
- "MUSR Score": "16.34 %",
348
- "MMLU-PRO Score": "47.59 %",
349
- "Model Architecture": "Qwen2ForCausalLM",
350
- "Model Precision": "bfloat16",
351
- "Model Parameters": "14.766B",
352
- "Chat Template Use": "Yes"
353
- },
354
- {
355
- "Model Rank": "3",
356
- "Icon": "ð¤",
357
- "Model Name": "wanlige/li-14b-v0.4-slerp0.1",
358
- "Average Score": "42.91 %",
359
- "IFEval Score": "79.23 %",
360
- "BBH Score": "50.88 %",
361
- "MATH Score": "53.32 %",
362
- "GPQA Score": "14.54 %",
363
- "MUSR Score": "11.75 %",
364
- "MMLU-PRO Score": "47.71 %",
365
- "Model Architecture": "Qwen2ForCausalLM",
366
- "Model Precision": "bfloat16",
367
- "Model Parameters": "14.766B",
368
- "Chat Template Use": "Yes"
369
- },
370
- {
371
- "Model Rank": "4",
372
- "Icon": "ð¬",
373
- "Model Name": "sthenno-com/miscii-14b-0218",
374
- "Average Score": "42.90 %",
375
- "IFEval Score": "76.56 %",
376
- "BBH Score": "50.64 %",
377
- "MATH Score": "51.44 %",
378
- "GPQA Score": "17.79 %",
379
- "MUSR Score": "13.21 %",
380
- "MMLU-PRO Score": "47.75 %",
381
- "Model Architecture": "Qwen2ForCausalLM",
382
- "Model Precision": "bfloat16",
383
- "Model Parameters": "14.766B",
384
- "Chat Template Use": "Yes"
385
- },
386
- {
387
- "Model Rank": "5",
388
- "Icon": "├░┬╢",
389
- "Model Name": "sthenno/tempesthenno-ppo-ckpt40",
390
- "Average Score": "42.74 %",
391
- "IFEval Score": "79.23 %",
392
- "BBH Score": "50.57 %",
393
- "MATH Score": "47.36 %",
394
- "GPQA Score": "17.00 %",
395
- "MUSR Score": "14.56 %",
396
- "MMLU-PRO Score": "47.69 %",
397
- "Model Architecture": "Qwen2ForCausalLM",
398
- "Model Precision": "bfloat16",
399
- "Model Parameters": "14.766B",
400
- "Chat Template Use": "Yes"
401
- },
402
- {
403
- "Model Rank": "6",
404
- "Icon": "ð¬",
405
- "Model Name": "tanliboy/lambda-qwen2.5-14b-dpo-test",
406
- "Average Score": "42.62 %",
407
- "IFEval Score": "82.31 %",
408
- "BBH Score": "48.45 %",
409
- "MATH Score": "54.61 %",
410
- "GPQA Score": "14.99 %",
411
- "MUSR Score": "12.59 %",
412
- "MMLU-PRO Score": "42.75 %",
413
- "Model Architecture": "Qwen2ForCausalLM",
414
- "Model Precision": "bfloat16",
415
- "Model Parameters": "14.77B",
416
- "Chat Template Use": "Yes"
417
- },
418
- {
419
- "Model Rank": "7",
420
- "Icon": "├░┬╢",
421
- "Model Name": "sthenno/tempesthenno-nuslerp-001",
422
- "Average Score": "42.59 %",
423
- "IFEval Score": "79.26 %",
424
- "BBH Score": "51.04 %",
425
- "MATH Score": "47.58 %",
426
- "GPQA Score": "16.44 %",
427
- "MUSR Score": "13.88 %",
428
- "MMLU-PRO Score": "47.30 %",
429
- "Model Architecture": "Qwen2ForCausalLM",
430
- "Model Precision": "bfloat16",
431
- "Model Parameters": "14.766B",
432
- "Chat Template Use": "Yes"
433
- },
434
- {
435
- "Model Rank": "8",
436
- "Icon": "ð¤",
437
- "Model Name": "YOYO-AI/Qwen2.5-14B-1M-YOYO-V3",
438
- "Average Score": "42.56 %",
439
- "IFEval Score": "83.98 %",
440
- "BBH Score": "49.47 %",
441
- "MATH Score": "53.55 %",
442
- "GPQA Score": "10.51 %",
443
- "MUSR Score": "11.10 %",
444
- "MMLU-PRO Score": "46.74 %",
445
- "Model Architecture": "Qwen2ForCausalLM",
446
- "Model Precision": "float16",
447
- "Model Parameters": "14.766B",
448
- "Chat Template Use": "Yes"
449
- },
450
- {
451
- "Model Rank": "9",
452
- "Icon": "├░┬╢",
453
- "Model Name": "Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4",
454
- "Average Score": "42.55 %",
455
- "IFEval Score": "82.92 %",
456
- "BBH Score": "48.05 %",
457
- "MATH Score": "54.23 %",
458
- "GPQA Score": "12.30 %",
459
- "MUSR Score": "13.15 %",
460
- "MMLU-PRO Score": "44.65 %",
461
- "Model Architecture": "Qwen2ForCausalLM",
462
- "Model Precision": "bfloat16",
463
- "Model Parameters": "14.77B",
464
- "Chat Template Use": "Yes"
465
- },
466
- {
467
- "Model Rank": "10",
468
- "Icon": "ð¤",
469
- "Model Name": "djuna/Q2.5-Veltha-14B",
470
- "Average Score": "42.52 %",
471
- "IFEval Score": "82.92 %",
472
- "BBH Score": "49.75 %",
473
- "MATH Score": "47.89 %",
474
- "GPQA Score": "14.54 %",
475
- "MUSR Score": "12.26 %",
476
- "MMLU-PRO Score": "47.76 %",
477
- "Model Architecture": "Qwen2ForCausalLM",
478
- "Model Precision": "bfloat16",
479
- "Model Parameters": "14.766B",
480
- "Chat Template Use": "Yes"
481
- },
482
- {
483
- "Model Rank": "11",
484
- "Icon": "├░┬╢",
485
- "Model Name": "arcee-ai/Virtuoso-Small-v2",
486
- "Average Score": "42.48 %",
487
- "IFEval Score": "82.73 %",
488
- "BBH Score": "50.95 %",
489
- "MATH Score": "46.60 %",
490
- "GPQA Score": "13.76 %",
491
- "MUSR Score": "14.28 %",
492
- "MMLU-PRO Score": "46.53 %",
493
- "Model Architecture": "Qwen2ForCausalLM",
494
- "Model Precision": "bfloat16",
495
- "Model Parameters": "14.766B",
496
- "Chat Template Use": "Yes"
497
- },
498
- {
499
- "Model Rank": "12",
500
- "Icon": "ð¤",
501
- "Model Name": "YOYO-AI/Qwen2.5-14B-YOYO-V4-p1",
502
- "Average Score": "42.46 %",
503
- "IFEval Score": "82.03 %",
504
- "BBH Score": "50.25 %",
505
- "MATH Score": "53.32 %",
506
- "GPQA Score": "12.75 %",
507
- "MUSR Score": "11.73 %",
508
- "MMLU-PRO Score": "44.67 %",
509
- "Model Architecture": "Qwen2ForCausalLM",
510
- "Model Precision": "float16",
511
- "Model Parameters": "14.766B",
512
- "Chat Template Use": "Yes"
513
- },
514
- {
515
- "Model Rank": "13",
516
- "Icon": "ð¬",
517
- "Model Name": "jpacifico/Chocolatine-14B-Instruct-DPO-v1.3",
518
- "Average Score": "42.42 %",
519
- "IFEval Score": "70.40 %",
520
- "BBH Score": "54.85 %",
521
- "MATH Score": "56.19 %",
522
- "GPQA Score": "12.19 %",
523
- "MUSR Score": "12.29 %",
524
- "MMLU-PRO Score": "48.60 %",
525
- "Model Architecture": "Phi3ForCausalLM",
526
- "Model Precision": "float16",
527
- "Model Parameters": "14.66B",
528
- "Chat Template Use": "Yes"
529
- },
530
- {
531
- "Model Rank": "14",
532
- "Icon": "ð¬",
533
- "Model Name": "sthenno-com/miscii-14b-1028",
534
- "Average Score": "42.38 %",
535
- "IFEval Score": "82.37 %",
536
- "BBH Score": "49.26 %",
537
- "MATH Score": "50.30 %",
538
- "GPQA Score": "14.21 %",
539
- "MUSR Score": "12.00 %",
540
- "MMLU-PRO Score": "46.14 %",
541
- "Model Architecture": "Qwen2ForCausalLM",
542
- "Model Precision": "bfloat16",
543
- "Model Parameters": "14.77B",
544
- "Chat Template Use": "Yes"
545
- },
546
- {
547
- "Model Rank": "15",
548
- "Icon": "ð¬",
549
- "Model Name": "sthenno-com/miscii-14b-1225",
550
- "Average Score": "42.35 %",
551
- "IFEval Score": "78.78 %",
552
- "BBH Score": "50.91 %",
553
- "MATH Score": "45.17 %",
554
- "GPQA Score": "17.00 %",
555
- "MUSR Score": "14.77 %",
556
- "MMLU-PRO Score": "47.46 %",
557
- "Model Architecture": "Qwen2ForCausalLM",
558
- "Model Precision": "bfloat16",
559
- "Model Parameters": "14.766B",
560
- "Chat Template Use": "Yes"
561
- },
562
- {
563
- "Model Rank": "16",
564
- "Icon": "├░┬╢",
565
- "Model Name": "prithivMLmods/Sombrero-Opus-14B-Elite5",
566
- "Average Score": "42.32 %",
567
- "IFEval Score": "78.81 %",
568
- "BBH Score": "50.17 %",
569
- "MATH Score": "53.55 %",
570
- "GPQA Score": "11.52 %",
571
- "MUSR Score": "13.22 %",
572
- "MMLU-PRO Score": "46.67 %",
573
- "Model Architecture": "Qwen2ForCausalLM",
574
- "Model Precision": "float16",
575
- "Model Parameters": "14.766B",
576
- "Chat Template Use": "Yes"
577
- },
578
- {
579
- "Model Rank": "17",
580
- "Icon": "ð¤",
581
- "Model Name": "Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v8",
582
- "Average Score": "42.26 %",
583
- "IFEval Score": "73.84 %",
584
- "BBH Score": "49.31 %",
585
- "MATH Score": "41.69 %",
586
- "GPQA Score": "18.23 %",
587
- "MUSR Score": "21.96 %",
588
- "MMLU-PRO Score": "48.50 %",
589
- "Model Architecture": "Qwen2ForCausalLM",
590
- "Model Precision": "bfloat16",
591
- "Model Parameters": "14.766B",
592
- "Chat Template Use": "No"
593
- },
594
- {
595
- "Model Rank": "18",
596
- "Icon": "├░┬╢",
597
- "Model Name": "prithivMLmods/Equuleus-Opus-14B-Exp",
598
- "Average Score": "42.20 %",
599
- "IFEval Score": "70.01 %",
600
- "BBH Score": "48.62 %",
601
- "MATH Score": "45.85 %",
602
- "GPQA Score": "18.23 %",
603
- "MUSR Score": "21.90 %",
604
- "MMLU-PRO Score": "48.60 %",
605
- "Model Architecture": "Qwen2ForCausalLM",
606
- "Model Precision": "bfloat16",
607
- "Model Parameters": "14.766B",
608
- "Chat Template Use": "No"
609
- },
610
- {
611
- "Model Rank": "19",
612
- "Icon": "├░┬╢",
613
- "Model Name": "rombodawg/Rombos-LLM-V2.6-Qwen-14b",
614
- "Average Score": "42.20 %",
615
- "IFEval Score": "84.32 %",
616
- "BBH Score": "49.28 %",
617
- "MATH Score": "52.11 %",
618
- "GPQA Score": "11.19 %",
619
- "MUSR Score": "12.29 %",
620
- "MMLU-PRO Score": "44.01 %",
621
- "Model Architecture": "Qwen2ForCausalLM",
622
- "Model Precision": "bfloat16",
623
- "Model Parameters": "14.77B",
624
- "Chat Template Use": "Yes"
625
- },
626
- {
627
- "Model Rank": "20",
628
- "Icon": "ð¤",
629
- "Model Name": "nbeerbower/EVA-abliterated-TIES-Qwen2.5-14B",
630
- "Average Score": "42.16 %",
631
- "IFEval Score": "78.36 %",
632
- "BBH Score": "48.52 %",
633
- "MATH Score": "50.45 %",
634
- "GPQA Score": "13.98 %",
635
- "MUSR Score": "14.88 %",
636
- "MMLU-PRO Score": "46.79 %",
637
- "Model Architecture": "Qwen2ForCausalLM",
638
- "Model Precision": "bfloat16",
639
- "Model Parameters": "14.77B",
640
- "Chat Template Use": "Yes"
641
- },
642
- {
643
- "Model Rank": "21",
644
- "Icon": "ð¤",
645
- "Model Name": "sometimesanotion/LamarckInfusion-14B-v1",
646
- "Average Score": "42.06 %",
647
- "IFEval Score": "71.98 %",
648
- "BBH Score": "50.35 %",
649
- "MATH Score": "41.69 %",
650
- "GPQA Score": "18.79 %",
651
- "MUSR Score": "20.90 %",
652
- "MMLU-PRO Score": "48.63 %",
653
- "Model Architecture": "Qwen2ForCausalLM",
654
- "Model Precision": "bfloat16",
655
- "Model Parameters": "14.766B",
656
- "Chat Template Use": "No"
657
- },
658
- {
659
- "Model Rank": "22",
660
- "Icon": "ð¤",
661
- "Model Name": "tensopolis/virtuoso-small-v2-tensopolis-v1",
662
- "Average Score": "41.99 %",
663
- "IFEval Score": "82.40 %",
664
- "BBH Score": "50.53 %",
665
- "MATH Score": "46.53 %",
666
- "GPQA Score": "12.53 %",
667
- "MUSR Score": "13.88 %",
668
- "MMLU-PRO Score": "46.07 %",
669
- "Model Architecture": "Qwen2ForCausalLM",
670
- "Model Precision": "bfloat16",
671
- "Model Parameters": "14.766B",
672
- "Chat Template Use": "Yes"
673
- },
674
- {
675
- "Model Rank": "23",
676
- "Icon": "ð¤",
677
- "Model Name": "Quazim0t0/Fugazi14b",
678
- "Average Score": "41.94 %",
679
- "IFEval Score": "69.98 %",
680
- "BBH Score": "56.09 %",
681
- "MATH Score": "46.53 %",
682
- "GPQA Score": "13.53 %",
683
- "MUSR Score": "16.42 %",
684
- "MMLU-PRO Score": "49.08 %",
685
- "Model Architecture": "LlamaForCausalLM",
686
- "Model Precision": "bfloat16",
687
- "Model Parameters": "14.66B",
688
- "Chat Template Use": "Yes"
689
- },
690
- {
691
- "Model Rank": "24",
692
- "Icon": "├░┬╢",
693
- "Model Name": "1024m/QWEN-14B-B100",
694
- "Average Score": "41.92 %",
695
- "IFEval Score": "77.62 %",
696
- "BBH Score": "49.78 %",
697
- "MATH Score": "54.38 %",
698
- "GPQA Score": "13.42 %",
699
- "MUSR Score": "9.88 %",
700
- "MMLU-PRO Score": "46.43 %",
701
- "Model Architecture": "Qwen2ForCausalLM",
702
- "Model Precision": "bfloat16",
703
- "Model Parameters": "14.77B",
704
- "Chat Template Use": "Yes"
705
- },
706
- {
707
- "Model Rank": "25",
708
- "Icon": "├░┬╢",
709
- "Model Name": "Sakalti/Saka-14B",
710
- "Average Score": "41.91 %",
711
- "IFEval Score": "71.74 %",
712
- "BBH Score": "49.72 %",
713
- "MATH Score": "40.94 %",
714
- "GPQA Score": "19.46 %",
715
- "MUSR Score": "20.74 %",
716
- "MMLU-PRO Score": "48.84 %",
717
- "Model Architecture": "Qwen2ForCausalLM",
718
- "Model Precision": "float16",
719
- "Model Parameters": "14.766B",
720
- "Chat Template Use": "No"
721
- },
722
- {
723
- "Model Rank": "26",
724
- "Icon": "├░┬╢",
725
- "Model Name": "prithivMLmods/Sombrero-Opus-14B-Elite6",
726
- "Average Score": "41.88 %",
727
- "IFEval Score": "72.26 %",
728
- "BBH Score": "49.60 %",
729
- "MATH Score": "40.79 %",
730
- "GPQA Score": "19.13 %",
731
- "MUSR Score": "20.74 %",
732
- "MMLU-PRO Score": "48.78 %",
733
- "Model Architecture": "Qwen2ForCausalLM",
734
- "Model Precision": "bfloat16",
735
- "Model Parameters": "14.766B",
736
- "Chat Template Use": "No"
737
- },
738
- {
739
- "Model Rank": "27",
740
- "Icon": "ð¤",
741
- "Model Name": "YOYO-AI/Qwen2.5-14B-YOYO-latest-V2",
742
- "Average Score": "41.85 %",
743
- "IFEval Score": "77.71 %",
744
- "BBH Score": "47.30 %",
745
- "MATH Score": "51.59 %",
746
- "GPQA Score": "13.87 %",
747
- "MUSR Score": "13.68 %",
748
- "MMLU-PRO Score": "46.93 %",
749
- "Model Architecture": "Qwen2ForCausalLM",
750
- "Model Precision": "float16",
751
- "Model Parameters": "14.766B",
752
- "Chat Template Use": "Yes"
753
- },
754
- {
755
- "Model Rank": "28",
756
- "Icon": "ð¬",
757
- "Model Name": "Tsunami-th/Tsunami-1.0-14B-Instruct",
758
- "Average Score": "41.84 %",
759
- "IFEval Score": "78.29 %",
760
- "BBH Score": "49.15 %",
761
- "MATH Score": "45.85 %",
762
- "GPQA Score": "14.21 %",
763
- "MUSR Score": "16.34 %",
764
- "MMLU-PRO Score": "47.21 %",
765
- "Model Architecture": "Qwen2ForCausalLM",
766
- "Model Precision": "bfloat16",
767
- "Model Parameters": "14.77B",
768
- "Chat Template Use": "Yes"
769
- },
770
- {
771
- "Model Rank": "29",
772
- "Icon": "├░┬╢",
773
- "Model Name": "sthenno/tempesthenno-kto-0205-ckpt80",
774
- "Average Score": "41.79 %",
775
- "IFEval Score": "80.54 %",
776
- "BBH Score": "50.64 %",
777
- "MATH Score": "45.92 %",
778
- "GPQA Score": "13.09 %",
779
- "MUSR Score": "12.93 %",
780
- "MMLU-PRO Score": "47.62 %",
781
- "Model Architecture": "Qwen2ForCausalLM",
782
- "Model Precision": "bfloat16",
783
- "Model Parameters": "14.766B",
784
- "Chat Template Use": "No"
785
- },
786
- {
787
- "Model Rank": "30",
788
- "Icon": "ð¤",
789
- "Model Name": "sometimesanotion/Lamarck-14B-v0.7-rc4",
790
- "Average Score": "41.79 %",
791
- "IFEval Score": "72.11 %",
792
- "BBH Score": "49.85 %",
793
- "MATH Score": "40.26 %",
794
- "GPQA Score": "18.57 %",
795
- "MUSR Score": "21.07 %",
796
- "MMLU-PRO Score": "48.89 %",
797
- "Model Architecture": "Qwen2ForCausalLM",
798
- "Model Precision": "bfloat16",
799
- "Model Parameters": "14.766B",
800
- "Chat Template Use": "No"
801
- },
802
- {
803
- "Model Rank": "31",
804
- "Icon": "├░┬╢",
805
- "Model Name": "prithivMLmods/Porpoise-Opus-14B-Exp",
806
- "Average Score": "41.77 %",
807
- "IFEval Score": "70.98 %",
808
- "BBH Score": "49.95 %",
809
- "MATH Score": "40.41 %",
810
- "GPQA Score": "19.13 %",
811
- "MUSR Score": "21.30 %",
812
- "MMLU-PRO Score": "48.85 %",
813
- "Model Architecture": "Qwen2ForCausalLM",
814
- "Model Precision": "bfloat16",
815
- "Model Parameters": "14.766B",
816
- "Chat Template Use": "No"
817
- },
818
- {
819
- "Model Rank": "32",
820
- "Icon": "ð¤",
821
- "Model Name": "CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES",
822
- "Average Score": "41.77 %",
823
- "IFEval Score": "82.40 %",
824
- "BBH Score": "48.20 %",
825
- "MATH Score": "53.17 %",
826
- "GPQA Score": "9.96 %",
827
- "MUSR Score": "12.65 %",
828
- "MMLU-PRO Score": "44.21 %",
829
- "Model Architecture": "Qwen2ForCausalLM",
830
- "Model Precision": "bfloat16",
831
- "Model Parameters": "14.77B",
832
- "Chat Template Use": "Yes"
833
- },
834
- {
835
- "rank": 1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
836
  "name": "suayptalha/Lamarckvergence-14B",
837
  "scores": {
838
  "average": 43.32,
@@ -847,10 +911,10 @@ benchmark_data = [
847
  "Chat_Template": "Yes"
848
  },
849
  "hf_url": "https://huggingface.co/suayptalha/Lamarckvergence-14B",
850
- "known_config": None
851
  },
852
  {
853
- "rank": 2,
854
  "name": "sthenno/tempesthenno-ppo-ckpt40",
855
  "scores": {
856
  "average": 42.74,
@@ -865,10 +929,10 @@ benchmark_data = [
865
  "Chat_Template": "Yes"
866
  },
867
  "hf_url": "https://huggingface.co/sthenno/tempesthenno-ppo-ckpt40",
868
- "known_config": None
869
  },
870
  {
871
- "rank": 3,
872
  "name": "tanliboy/lambda-qwen2.5-14b-dpo-test",
873
  "scores": {
874
  "average": 42.62,
@@ -883,10 +947,10 @@ benchmark_data = [
883
  "Chat_Template": "Yes"
884
  },
885
  "hf_url": "https://huggingface.co/tanliboy/lambda-qwen2.5-14b-dpo-test",
886
- "known_config": None
887
  },
888
  {
889
- "rank": 4,
890
  "name": "sthenno/tempesthenno-nuslerp-001",
891
  "scores": {
892
  "average": 42.59,
@@ -901,10 +965,10 @@ benchmark_data = [
901
  "Chat_Template": "Yes"
902
  },
903
  "hf_url": "https://huggingface.co/sthenno/tempesthenno-nuslerp-001",
904
- "known_config": None
905
  },
906
  {
907
- "rank": 5,
908
  "name": "Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4",
909
  "scores": {
910
  "average": 42.55,
@@ -919,10 +983,10 @@ benchmark_data = [
919
  "Chat_Template": "Yes"
920
  },
921
  "hf_url": "https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4",
922
- "known_config": None
923
  },
924
  {
925
- "rank": 6,
926
  "name": "djuna/Q2.5-Veltha-14B",
927
  "scores": {
928
  "average": 42.52,
@@ -937,10 +1001,10 @@ benchmark_data = [
937
  "Chat_Template": "Yes"
938
  },
939
  "hf_url": "https://huggingface.co/djuna/Q2.5-Veltha-14B",
940
- "known_config": None
941
  },
942
  {
943
- "rank": 7,
944
  "name": "arcee-ai/Virtuoso-Small-v2",
945
  "scores": {
946
  "average": 42.48,
@@ -955,10 +1019,10 @@ benchmark_data = [
955
  "Chat_Template": "Yes"
956
  },
957
  "hf_url": "https://huggingface.co/arcee-ai/Virtuoso-Small-v2",
958
- "known_config": None
959
  },
960
  {
961
- "rank": 8,
962
  "name": "jpacifico/Chocolatine-14B-Instruct-DPO-v1.3",
963
  "scores": {
964
  "average": 42.42,
@@ -973,10 +1037,10 @@ benchmark_data = [
973
  "Chat_Template": "Yes"
974
  },
975
  "hf_url": "https://huggingface.co/jpacifico/Chocolatine-14B-Instruct-DPO-v1.3",
976
- "known_config": None
977
  },
978
  {
979
- "rank": 9,
980
  "name": "sthenno-com/miscii-14b-1028",
981
  "scores": {
982
  "average": 42.38,
@@ -991,10 +1055,10 @@ benchmark_data = [
991
  "Chat_Template": "Yes"
992
  },
993
  "hf_url": "https://huggingface.co/sthenno-com/miscii-14b-1028",
994
- "known_config": None
995
  },
996
  {
997
- "rank": 10,
998
  "name": "sthenno-com/miscii-14b-1225",
999
  "scores": {
1000
  "average": 42.35,
@@ -1009,10 +1073,10 @@ benchmark_data = [
1009
  "Chat_Template": "Yes"
1010
  },
1011
  "hf_url": "https://huggingface.co/sthenno-com/miscii-14b-1225",
1012
- "known_config": None
1013
  },
1014
  {
1015
- "rank": 11,
1016
  "name": "tensopolis/virtuoso-small-v2-tensopolis-v1",
1017
  "scores": {
1018
  "average": 42.34,
@@ -1027,10 +1091,10 @@ benchmark_data = [
1027
  "Chat_Template": "Yes"
1028
  },
1029
  "hf_url": "https://huggingface.co/tensopolis/virtuoso-small-v2-tensopolis-v1",
1030
- "known_config": None
1031
  },
1032
  {
1033
- "rank": 12,
1034
  "name": "rombodawg/Rombos-LLM-V2.6-Qwen-14b",
1035
  "scores": {
1036
  "average": 42.2,
@@ -1045,10 +1109,10 @@ benchmark_data = [
1045
  "Chat_Template": "Yes"
1046
  },
1047
  "hf_url": "https://huggingface.co/rombodawg/Rombos-LLM-V2.6-Qwen-14b",
1048
- "known_config": None
1049
  },
1050
  {
1051
- "rank": 13,
1052
  "name": "1024m/QWEN-14B-B100",
1053
  "scores": {
1054
  "average": 41.92,
@@ -1063,10 +1127,10 @@ benchmark_data = [
1063
  "Chat_Template": "Yes"
1064
  },
1065
  "hf_url": "https://huggingface.co/1024m/QWEN-14B-B100",
1066
- "known_config": None
1067
  },
1068
  {
1069
- "rank": 14,
1070
  "name": "Sakalti/Saka-14B",
1071
  "scores": {
1072
  "average": 41.91,
@@ -1081,10 +1145,10 @@ benchmark_data = [
1081
  "Chat_Template": "No"
1082
  },
1083
  "hf_url": "https://huggingface.co/Sakalti/Saka-14B",
1084
- "known_config": None
1085
  },
1086
  {
1087
- "rank": 15,
1088
  "name": "Tsunami-th/Tsunami-1.0-14B-Instruct",
1089
  "scores": {
1090
  "average": 41.84,
@@ -1099,10 +1163,10 @@ benchmark_data = [
1099
  "Chat_Template": "Yes"
1100
  },
1101
  "hf_url": "https://huggingface.co/Tsunami-th/Tsunami-1.0-14B-Instruct",
1102
- "known_config": None
1103
  },
1104
  {
1105
- "rank": 16,
1106
  "name": "sthenno/tempesthenno-kto-0205-ckpt80",
1107
  "scores": {
1108
  "average": 41.79,
@@ -1117,10 +1181,10 @@ benchmark_data = [
1117
  "Chat_Template": "No"
1118
  },
1119
  "hf_url": "https://huggingface.co/sthenno/tempesthenno-kto-0205-ckpt80",
1120
- "known_config": None
1121
  },
1122
  {
1123
- "rank": 17,
1124
  "name": "sometimesanotion/Lamarck-14B-v0.7-rc4",
1125
  "scores": {
1126
  "average": 41.79,
@@ -1135,10 +1199,10 @@ benchmark_data = [
1135
  "Chat_Template": "No"
1136
  },
1137
  "hf_url": "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7-rc4",
1138
- "known_config": None
1139
  },
1140
  {
1141
- "rank": 18,
1142
  "name": "CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES",
1143
  "scores": {
1144
  "average": 41.77,
@@ -1153,10 +1217,10 @@ benchmark_data = [
1153
  "Chat_Template": "Yes"
1154
  },
1155
  "hf_url": "https://huggingface.co/CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES",
1156
- "known_config": None
1157
  },
1158
  {
1159
- "rank": 19,
1160
  "name": "suayptalha/Luminis-phi-4",
1161
  "scores": {
1162
  "average": 41.76,
@@ -1171,10 +1235,10 @@ benchmark_data = [
1171
  "Chat_Template": "Yes"
1172
  },
1173
  "hf_url": "https://huggingface.co/suayptalha/Luminis-phi-4",
1174
- "known_config": None
1175
  },
1176
  {
1177
- "rank": 20,
1178
  "name": "huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2",
1179
  "scores": {
1180
  "average": 41.75,
@@ -1189,10 +1253,10 @@ benchmark_data = [
1189
  "Chat_Template": "Yes"
1190
  },
1191
  "hf_url": "https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2",
1192
- "known_config": None
1193
  },
1194
  {
1195
- "rank": 21,
1196
  "name": "djuna/Q2.5-Veltha-14B-0.5",
1197
  "scores": {
1198
  "average": 41.61,
@@ -1207,10 +1271,10 @@ benchmark_data = [
1207
  "Chat_Template": "Yes"
1208
  },
1209
  "hf_url": "https://huggingface.co/djuna/Q2.5-Veltha-14B-0.5",
1210
- "known_config": None
1211
  },
1212
  {
1213
- "rank": 23,
1214
  "name": "Qwen/Qwen2.5-14B-Instruct-1M",
1215
  "scores": {
1216
  "average": 41.56,
@@ -1225,10 +1289,10 @@ benchmark_data = [
1225
  "Chat_Template": "Yes"
1226
  },
1227
  "hf_url": "https://huggingface.co/Qwen/Qwen2.5-14B-Instruct-1M",
1228
- "known_config": None
1229
  },
1230
  {
1231
- "rank": 22,
1232
  "name": "notbdq/Qwen2.5-14B-Instruct-1M-GRPO-Reasoning",
1233
  "scores": {
1234
  "average": 41.56,
@@ -1243,10 +1307,10 @@ benchmark_data = [
1243
  "Chat_Template": "Yes"
1244
  },
1245
  "hf_url": "https://huggingface.co/notbdq/Qwen2.5-14B-Instruct-1M-GRPO-Reasoning",
1246
- "known_config": None
1247
  },
1248
  {
1249
- "rank": 24,
1250
  "name": "sometimesanotion/Qwenvergence-14B-v11",
1251
  "scores": {
1252
  "average": 41.52,
@@ -1261,10 +1325,10 @@ benchmark_data = [
1261
  "Chat_Template": "No"
1262
  },
1263
  "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v11",
1264
- "known_config": None
1265
  },
1266
  {
1267
- "rank": 25,
1268
  "name": "sometimesanotion/Qwenvergence-14B-v10",
1269
  "scores": {
1270
  "average": 41.48,
@@ -1279,10 +1343,10 @@ benchmark_data = [
1279
  "Chat_Template": "No"
1280
  },
1281
  "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v10",
1282
- "known_config": None
1283
  },
1284
  {
1285
- "rank": 26,
1286
  "name": "CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES",
1287
  "scores": {
1288
  "average": 41.47,
@@ -1297,10 +1361,10 @@ benchmark_data = [
1297
  "Chat_Template": "Yes"
1298
  },
1299
  "hf_url": "https://huggingface.co/CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES",
1300
- "known_config": None
1301
  },
1302
  {
1303
- "rank": 27,
1304
  "name": "RDson/WomboCombo-R1-Coder-14B-Preview",
1305
  "scores": {
1306
  "average": 41.46,
@@ -1315,10 +1379,10 @@ benchmark_data = [
1315
  "Chat_Template": "Yes"
1316
  },
1317
  "hf_url": "https://huggingface.co/RDson/WomboCombo-R1-Coder-14B-Preview",
1318
- "known_config": None
1319
  },
1320
  {
1321
- "rank": 28,
1322
  "name": "jpacifico/Chocolatine-2-14B-Instruct-v2.0b3",
1323
  "scores": {
1324
  "average": 41.43,
@@ -1333,10 +1397,10 @@ benchmark_data = [
1333
  "Chat_Template": "No"
1334
  },
1335
  "hf_url": "https://huggingface.co/jpacifico/Chocolatine-2-14B-Instruct-v2.0b3",
1336
- "known_config": None
1337
  },
1338
  {
1339
- "rank": 29,
1340
  "name": "Quazim0t0/Nova-14b-sce",
1341
  "scores": {
1342
  "average": 41.41,
@@ -1351,10 +1415,10 @@ benchmark_data = [
1351
  "Chat_Template": "Yes"
1352
  },
1353
  "hf_url": "https://huggingface.co/Quazim0t0/Nova-14b-sce",
1354
- "known_config": None
1355
  },
1356
  {
1357
- "rank": 30,
1358
  "name": "v000000/Qwen2.5-14B-Gutenberg-Instruct-Slerpeno",
1359
  "scores": {
1360
  "average": 41.36,
@@ -1369,10 +1433,10 @@ benchmark_data = [
1369
  "Chat_Template": "Yes"
1370
  },
1371
  "hf_url": "https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-Instruct-Slerpeno",
1372
- "known_config": None
1373
  },
1374
  {
1375
- "rank": 31,
1376
  "name": "Quazim0t0/NovaScotia-14b-stock",
1377
  "scores": {
1378
  "average": 41.35,
@@ -1387,10 +1451,10 @@ benchmark_data = [
1387
  "Chat_Template": "Yes"
1388
  },
1389
  "hf_url": "https://huggingface.co/Quazim0t0/NovaScotia-14b-stock",
1390
- "known_config": None
1391
  },
1392
  {
1393
- "rank": 32,
1394
  "name": "Quazim0t0/ODB-14b-sce",
1395
  "scores": {
1396
  "average": 41.34,
@@ -1405,10 +1469,10 @@ benchmark_data = [
1405
  "Chat_Template": "Yes"
1406
  },
1407
  "hf_url": "https://huggingface.co/Quazim0t0/ODB-14b-sce",
1408
- "known_config": None
1409
  },
1410
  {
1411
- "rank": 54,
1412
  "name": "LightningRodLabs/Flashlight-v1.1",
1413
  "scores": {
1414
  "average": 40.99,
@@ -1423,10 +1487,10 @@ benchmark_data = [
1423
  "Chat_Template": "Yes"
1424
  },
1425
  "hf_url": "https://huggingface.co/LightningRodLabs/Flashlight-v1.1",
1426
- "known_config": None
1427
  },
1428
  {
1429
- "rank": 55,
1430
  "name": "Quazim0t0/Mithril-14B-sce",
1431
  "scores": {
1432
  "average": 40.98,
@@ -1441,10 +1505,10 @@ benchmark_data = [
1441
  "Chat_Template": "Yes"
1442
  },
1443
  "hf_url": "https://huggingface.co/Quazim0t0/Mithril-14B-sce",
1444
- "known_config": None
1445
  },
1446
  {
1447
- "rank": 56,
1448
  "name": "Sakalti/ultiima-14B-v0.2",
1449
  "scores": {
1450
  "average": 40.96,
@@ -1459,10 +1523,10 @@ benchmark_data = [
1459
  "Chat_Template": "No"
1460
  },
1461
  "hf_url": "https://huggingface.co/Sakalti/ultiima-14B-v0.2",
1462
- "known_config": None
1463
  },
1464
  {
1465
- "rank": 57,
1466
  "name": "bunnycore/Phi-4-ReasoningRP",
1467
  "scores": {
1468
  "average": 40.95,
@@ -1477,10 +1541,10 @@ benchmark_data = [
1477
  "Chat_Template": "Yes"
1478
  },
1479
  "hf_url": "https://huggingface.co/bunnycore/Phi-4-ReasoningRP",
1480
- "known_config": None
1481
  },
1482
  {
1483
- "rank": 58,
1484
  "name": "dwikitheduck/gen-inst-1",
1485
  "scores": {
1486
  "average": 40.88,
@@ -1495,10 +1559,10 @@ benchmark_data = [
1495
  "Chat_Template": "Yes"
1496
  },
1497
  "hf_url": "https://huggingface.co/dwikitheduck/gen-inst-1",
1498
- "known_config": None
1499
  },
1500
  {
1501
- "rank": 59,
1502
  "name": "v000000/Qwen2.5-14B-Gutenberg-1e-Delta",
1503
  "scores": {
1504
  "average": 40.88,
@@ -1513,7 +1577,7 @@ benchmark_data = [
1513
  "Chat_Template": "Yes"
1514
  },
1515
  "hf_url": "https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-1e-Delta",
1516
- "known_config": None
1517
  },
1518
  {
1519
  "rank": 60,
@@ -1531,10 +1595,10 @@ benchmark_data = [
1531
  "Chat_Template": "No"
1532
  },
1533
  "hf_url": "https://huggingface.co/hotmailuser/QwenSlerp2-14B",
1534
- "known_config": None
1535
  },
1536
  {
1537
- "rank": 61,
1538
  "name": "Quazim0t0/Loke-14B-sce",
1539
  "scores": {
1540
  "average": 40.86,
@@ -1549,10 +1613,10 @@ benchmark_data = [
1549
  "Chat_Template": "Yes"
1550
  },
1551
  "hf_url": "https://huggingface.co/Quazim0t0/Loke-14B-sce",
1552
- "known_config": None
1553
  },
1554
  {
1555
- "rank": 62,
1556
  "name": "Quazim0t0/mosaic-14b-sce",
1557
  "scores": {
1558
  "average": 40.83,
@@ -1567,10 +1631,10 @@ benchmark_data = [
1567
  "Chat_Template": "Yes"
1568
  },
1569
  "hf_url": "https://huggingface.co/Quazim0t0/mosaic-14b-sce",
1570
- "known_config": None
1571
  },
1572
  {
1573
- "rank": 63,
1574
  "name": "bunnycore/Phi-4-Model-Stock",
1575
  "scores": {
1576
  "average": 40.79,
@@ -1585,10 +1649,10 @@ benchmark_data = [
1585
  "Chat_Template": "Yes"
1586
  },
1587
  "hf_url": "https://huggingface.co/bunnycore/Phi-4-Model-Stock",
1588
- "known_config": None
1589
  },
1590
  {
1591
- "rank": 64,
1592
  "name": "unsloth/phi-4",
1593
  "scores": {
1594
  "average": 40.73,
@@ -1603,10 +1667,10 @@ benchmark_data = [
1603
  "Chat_Template": "Yes"
1604
  },
1605
  "hf_url": "https://huggingface.co/unsloth/phi-4",
1606
- "known_config": None
1607
  },
1608
  {
1609
- "rank": 65,
1610
  "name": "pankajmathur/orca_mini_phi-4",
1611
  "scores": {
1612
  "average": 40.68,
@@ -1621,10 +1685,10 @@ benchmark_data = [
1621
  "Chat_Template": "Yes"
1622
  },
1623
  "hf_url": "https://huggingface.co/pankajmathur/orca_mini_phi-4",
1624
- "known_config": None
1625
  },
1626
  {
1627
- "rank": 66,
1628
  "name": "pankajmathur/orca_mini_v9_2_14B",
1629
  "scores": {
1630
  "average": 40.68,
@@ -1639,10 +1703,10 @@ benchmark_data = [
1639
  "Chat_Template": "Yes"
1640
  },
1641
  "hf_url": "https://huggingface.co/pankajmathur/orca_mini_v9_2_14B",
1642
- "known_config": None
1643
  },
1644
  {
1645
- "rank": 67,
1646
  "name": "sometimesanotion/Lamarck-14B-v0.6-model_stock",
1647
  "scores": {
1648
  "average": 40.68,
@@ -1657,10 +1721,10 @@ benchmark_data = [
1657
  "Chat_Template": "No"
1658
  },
1659
  "hf_url": "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6-model_stock",
1660
- "known_config": None
1661
  },
1662
  {
1663
- "rank": 68,
1664
  "name": "sometimesanotion/Qwenvergence-14B-v0.6-004-model_stock",
1665
  "scores": {
1666
  "average": 40.6,
@@ -1675,10 +1739,10 @@ benchmark_data = [
1675
  "Chat_Template": "No"
1676
  },
1677
  "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v0.6-004-model_stock",
1678
- "known_config": None
1679
  },
1680
  {
1681
- "rank": 69,
1682
  "name": "Quazim0t0/Oasis-14B-ties",
1683
  "scores": {
1684
  "average": 40.59,
@@ -1693,10 +1757,10 @@ benchmark_data = [
1693
  "Chat_Template": "Yes"
1694
  },
1695
  "hf_url": "https://huggingface.co/Quazim0t0/Oasis-14B-ties",
1696
- "known_config": None
1697
  },
1698
  {
1699
- "rank": 70,
1700
  "name": "LightningRodLabs/Flashlight-v1.0",
1701
  "scores": {
1702
  "average": 40.57,
@@ -1711,10 +1775,10 @@ benchmark_data = [
1711
  "Chat_Template": "Yes"
1712
  },
1713
  "hf_url": "https://huggingface.co/LightningRodLabs/Flashlight-v1.0",
1714
- "known_config": None
1715
  },
1716
  {
1717
- "rank": 71,
1718
  "name": "arcee-ai/Virtuoso-Small",
1719
  "scores": {
1720
  "average": 40.54,
@@ -1729,10 +1793,10 @@ benchmark_data = [
1729
  "Chat_Template": "Yes"
1730
  },
1731
  "hf_url": "https://huggingface.co/arcee-ai/Virtuoso-Small",
1732
- "known_config": None
1733
  },
1734
  {
1735
- "rank": 72,
1736
  "name": "Quazim0t0/GuiltySpark-14B-ties",
1737
  "scores": {
1738
  "average": 40.52,
@@ -1747,10 +1811,10 @@ benchmark_data = [
1747
  "Chat_Template": "Yes"
1748
  },
1749
  "hf_url": "https://huggingface.co/Quazim0t0/GuiltySpark-14B-ties",
1750
- "known_config": None
1751
  },
1752
  {
1753
- "rank": 73,
1754
  "name": "ozone-ai/0x-lite",
1755
  "scores": {
1756
  "average": 40.48,
@@ -1765,10 +1829,10 @@ benchmark_data = [
1765
  "Chat_Template": "Yes"
1766
  },
1767
  "hf_url": "https://huggingface.co/ozone-ai/0x-lite",
1768
- "known_config": None
1769
  },
1770
  {
1771
- "rank": 74,
1772
  "name": "Quazim0t0/Casa-14b-sce",
1773
  "scores": {
1774
  "average": 40.41,
@@ -1783,10 +1847,10 @@ benchmark_data = [
1783
  "Chat_Template": "Yes"
1784
  },
1785
  "hf_url": "https://huggingface.co/Quazim0t0/Casa-14b-sce",
1786
- "known_config": None
1787
  },
1788
  {
1789
- "rank": 75,
1790
  "name": "Sakalti/ultiima-14B-v0.3",
1791
  "scores": {
1792
  "average": 40.38,
@@ -1801,10 +1865,10 @@ benchmark_data = [
1801
  "Chat_Template": "No"
1802
  },
1803
  "hf_url": "https://huggingface.co/Sakalti/ultiima-14B-v0.3",
1804
- "known_config": None
1805
  },
1806
  {
1807
- "rank": 76,
1808
  "name": "ehristoforu/fp4-14b-v1-fix",
1809
  "scores": {
1810
  "average": 40.37,
@@ -1819,10 +1883,10 @@ benchmark_data = [
1819
  "Chat_Template": "Yes"
1820
  },
1821
  "hf_url": "https://huggingface.co/ehristoforu/fp4-14b-v1-fix",
1822
- "known_config": None
1823
  },
1824
  {
1825
- "rank": 77,
1826
  "name": "FINGU-AI/Chocolatine-Fusion-14B",
1827
  "scores": {
1828
  "average": 40.36,
@@ -1837,10 +1901,10 @@ benchmark_data = [
1837
  "Chat_Template": "No"
1838
  },
1839
  "hf_url": "https://huggingface.co/FINGU-AI/Chocolatine-Fusion-14B",
1840
- "known_config": None
1841
  },
1842
  {
1843
- "rank": 78,
1844
  "name": "hotmailuser/QwenSlerp-14B",
1845
  "scores": {
1846
  "average": 40.35,
@@ -1855,10 +1919,10 @@ benchmark_data = [
1855
  "Chat_Template": "No"
1856
  },
1857
  "hf_url": "https://huggingface.co/hotmailuser/QwenSlerp-14B",
1858
- "known_config": None
1859
  },
1860
  {
1861
- "rank": 79,
1862
  "name": "Triangle104/Robo-Gutenberg_V1.0",
1863
  "scores": {
1864
  "average": 40.35,
@@ -1873,10 +1937,10 @@ benchmark_data = [
1873
  "Chat_Template": "No"
1874
  },
1875
  "hf_url": "https://huggingface.co/Triangle104/Robo-Gutenberg_V1.0",
1876
- "known_config": None
1877
  },
1878
  {
1879
- "rank": 80,
1880
  "name": "Quazim0t0/Adamant-14B-sce",
1881
  "scores": {
1882
  "average": 40.32,
@@ -1891,10 +1955,10 @@ benchmark_data = [
1891
  "Chat_Template": "Yes"
1892
  },
1893
  "hf_url": "https://huggingface.co/Quazim0t0/Adamant-14B-sce",
1894
- "known_config": None
1895
  },
1896
  {
1897
- "rank": 81,
1898
  "name": "Quazim0t0/Phi4Basis-14B-sce",
1899
  "scores": {
1900
  "average": 40.31,
@@ -1909,10 +1973,10 @@ benchmark_data = [
1909
  "Chat_Template": "Yes"
1910
  },
1911
  "hf_url": "https://huggingface.co/Quazim0t0/Phi4Basis-14B-sce",
1912
- "known_config": None
1913
  },
1914
  {
1915
- "rank": 82,
1916
  "name": "Quazim0t0/bloom-14b-stock",
1917
  "scores": {
1918
  "average": 40.29,
@@ -1927,10 +1991,10 @@ benchmark_data = [
1927
  "Chat_Template": "Yes"
1928
  },
1929
  "hf_url": "https://huggingface.co/Quazim0t0/bloom-14b-stock",
1930
- "known_config": None
1931
  },
1932
  {
1933
- "rank": 83,
1934
  "name": "sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-Prose01",
1935
  "scores": {
1936
  "average": 40.28,
@@ -1945,10 +2009,10 @@ benchmark_data = [
1945
  "Chat_Template": "No"
1946
  },
1947
  "hf_url": "https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-Prose01",
1948
- "known_config": None
1949
  },
1950
  {
1951
- "rank": 84,
1952
  "name": "Quazim0t0/Halo-14B-sce",
1953
  "scores": {
1954
  "average": 40.26,
@@ -1963,10 +2027,10 @@ benchmark_data = [
1963
  "Chat_Template": "Yes"
1964
  },
1965
  "hf_url": "https://huggingface.co/Quazim0t0/Halo-14B-sce",
1966
- "known_config": None
1967
  },
1968
  {
1969
- "rank": 85,
1970
  "name": "prithivMLmods/Calcium-Opus-14B-Elite2",
1971
  "scores": {
1972
  "average": 40.25,
@@ -1981,10 +2045,10 @@ benchmark_data = [
1981
  "Chat_Template": "No"
1982
  },
1983
  "hf_url": "https://huggingface.co/prithivMLmods/Calcium-Opus-14B-Elite2",
1984
- "known_config": None
1985
  },
1986
  {
1987
- "rank": 86,
1988
  "name": "SicariusSicariiStuff/Impish_QWEN_14B-1M",
1989
  "scores": {
1990
  "average": 40.24,
@@ -1999,10 +2063,10 @@ benchmark_data = [
1999
  "Chat_Template": "Yes"
2000
  },
2001
  "hf_url": "https://huggingface.co/SicariusSicariiStuff/Impish_QWEN_14B-1M",
2002
- "known_config": None
2003
  },
2004
  {
2005
- "rank": 87,
2006
  "name": "bunnycore/Phi-4-Stock-Ex",
2007
  "scores": {
2008
  "average": 40.22,
@@ -2017,10 +2081,10 @@ benchmark_data = [
2017
  "Chat_Template": "Yes"
2018
  },
2019
  "hf_url": "https://huggingface.co/bunnycore/Phi-4-Stock-Ex",
2020
- "known_config": None
2021
  },
2022
  {
2023
- "rank": 88,
2024
  "name": "sometimesanotion/Qwenvergence-14B-qv256",
2025
  "scores": {
2026
  "average": 40.12,
@@ -2035,10 +2099,10 @@ benchmark_data = [
2035
  "Chat_Template": "No"
2036
  },
2037
  "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-qv256",
2038
- "known_config": None
2039
  },
2040
  {
2041
- "rank": 89,
2042
  "name": "tensopolis/virtuoso-small-tensopolis-v2",
2043
  "scores": {
2044
  "average": 40.11,
@@ -2053,12 +2117,13 @@ benchmark_data = [
2053
  "Chat_Template": "Yes"
2054
  },
2055
  "hf_url": "https://huggingface.co/tensopolis/virtuoso-small-tensopolis-v2",
2056
- "known_config": None
2057
  }
2058
  ]
2059
 
2060
 
2061
 
 
2062
  def snippet_scrape_model_page(url):
2063
  """
2064
  Equivalent scraping function for the larger dataset
 
319
  # --------------------------------------------------------------------
320
  # This is your larger dataset, rank = 44..105
321
  benchmark_data = [
322
+ {
323
+ "rank": 1,
324
+ "name": "wanlige/li-14b-v0.4",
325
+ "scores": {
326
+ "average": 43.66,
327
+ "IFEval": 81.33,
328
+ "BBH": 50.38,
329
+ "MATH": 55.74,
330
+ "GPQA": 11.86,
331
+ "MUSR": 16.35,
332
+ "MMLU_PRO": 46.30,
333
+ "Architecture": "Qwen2ForCausalLM",
334
+ "Parameters": "14.77B",
335
+ "Chat_Template": "Yes"
336
+ },
337
+ "hf_url": "https://huggingface.co/wanlige/li-14b-v0.4",
338
+ "known_config": null
339
+ },
340
+ {
341
+ "rank": 2,
342
+ "name": "suayptalha/Lamarckvergence-14B",
343
+ "scores": {
344
+ "average": 43.32,
345
+ "IFEval": 76.56,
346
+ "BBH": 50.33,
347
+ "MATH": 54.0,
348
+ "GPQA": 15.1,
349
+ "MUSR": 16.34,
350
+ "MMLU_PRO": 47.59,
351
+ "Architecture": "Qwen2ForCausalLM",
352
+ "Parameters": "14.766B",
353
+ "Chat_Template": "Yes"
354
+ },
355
+ "hf_url": "https://huggingface.co/suayptalha/Lamarckvergence-14B",
356
+ "known_config": null
357
+ },
358
+ {
359
+ "rank": 3,
360
+ "name": "wanlige/li-14b-v0.4-slerp0.1",
361
+ "scores": {
362
+ "average": 42.91,
363
+ "IFEval": 79.23,
364
+ "BBH": 50.88,
365
+ "MATH": 53.32,
366
+ "GPQA": 14.54,
367
+ "MUSR": 11.75,
368
+ "MMLU_PRO": 47.71,
369
+ "Architecture": "Qwen2ForCausalLM",
370
+ "Parameters": "14.766B",
371
+ "Chat_Template": "Yes"
372
+ },
373
+ "hf_url": "https://huggingface.co/wanlige/li-14b-v0.4-slerp0.1",
374
+ "known_config": null
375
+ },
376
+ {
377
+ "rank": 4,
378
+ "name": "sthenno-com/miscii-14b-0218",
379
+ "scores": {
380
+ "average": 42.90,
381
+ "IFEval": 76.56,
382
+ "BBH": 50.64,
383
+ "MATH": 51.44,
384
+ "GPQA": 17.79,
385
+ "MUSR": 13.21,
386
+ "MMLU_PRO": 47.75,
387
+ "Architecture": "Qwen2ForCausalLM",
388
+ "Parameters": "14.766B",
389
+ "Chat_Template": "Yes"
390
+ },
391
+ "hf_url": "https://huggingface.co/sthenno-com/miscii-14b-0218",
392
+ "known_config": null
393
+ },
394
+ {
395
+ "rank": 5,
396
+ "name": "sthenno/tempesthenno-ppo-ckpt40",
397
+ "scores": {
398
+ "average": 42.74,
399
+ "IFEval": 79.23,
400
+ "BBH": 50.57,
401
+ "MATH": 47.36,
402
+ "GPQA": 17.0,
403
+ "MUSR": 14.56,
404
+ "MMLU_PRO": 47.69,
405
+ "Architecture": "Qwen2ForCausalLM",
406
+ "Parameters": "14.766B",
407
+ "Chat_Template": "Yes"
408
+ },
409
+ "hf_url": "https://huggingface.co/sthenno/tempesthenno-ppo-ckpt40",
410
+ "known_config": null
411
+ },
412
+ {
413
+ "rank": 6,
414
+ "name": "tanliboy/lambda-qwen2.5-14b-dpo-test",
415
+ "scores": {
416
+ "average": 42.62,
417
+ "IFEval": 82.31,
418
+ "BBH": 48.45,
419
+ "MATH": 54.61,
420
+ "GPQA": 14.99,
421
+ "MUSR": 12.59,
422
+ "MMLU_PRO": 42.75,
423
+ "Architecture": "Qwen2ForCausalLM",
424
+ "Parameters": "14.77B",
425
+ "Chat_Template": "Yes"
426
+ },
427
+ "hf_url": "https://huggingface.co/tanliboy/lambda-qwen2.5-14b-dpo-test",
428
+ "known_config": null
429
+ },
430
+ {
431
+ "rank": 7,
432
+ "name": "sthenno/tempesthenno-nuslerp-001",
433
+ "scores": {
434
+ "average": 42.59,
435
+ "IFEval": 79.26,
436
+ "BBH": 51.04,
437
+ "MATH": 47.58,
438
+ "GPQA": 16.44,
439
+ "MUSR": 13.88,
440
+ "MMLU_PRO": 47.30,
441
+ "Architecture": "Qwen2ForCausalLM",
442
+ "Parameters": "14.766B",
443
+ "Chat_Template": "Yes"
444
+ },
445
+ "hf_url": "https://huggingface.co/sthenno/tempesthenno-nuslerp-001",
446
+ "known_config": null
447
+ },
448
+ {
449
+ "rank": 8,
450
+ "name": "YOYO-AI/Qwen2.5-14B-1M-YOYO-V3",
451
+ "scores": {
452
+ "average": 42.56,
453
+ "IFEval": 83.98,
454
+ "BBH": 49.47,
455
+ "MATH": 53.55,
456
+ "GPQA": 10.51,
457
+ "MUSR": 11.10,
458
+ "MMLU_PRO": 46.74,
459
+ "Architecture": "Qwen2ForCausalLM",
460
+ "Parameters": "14.766B",
461
+ "Chat_Template": "Yes"
462
+ },
463
+ "hf_url": "https://huggingface.co/YOYO-AI/Qwen2.5-14B-1M-YOYO-V3",
464
+ "known_config": null
465
+ },
466
+ {
467
+ "rank": 9,
468
+ "name": "Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4",
469
+ "scores": {
470
+ "average": 42.55,
471
+ "IFEval": 82.92,
472
+ "BBH": 48.05,
473
+ "MATH": 54.23,
474
+ "GPQA": 12.30,
475
+ "MUSR": 13.15,
476
+ "MMLU_PRO": 44.65,
477
+ "Architecture": "Qwen2ForCausalLM",
478
+ "Parameters": "14.77B",
479
+ "Chat_Template": "Yes"
480
+ },
481
+ "hf_url": "https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4",
482
+ "known_config": null
483
+ },
484
+ {
485
+ "rank": 10,
486
+ "name": "djuna/Q2.5-Veltha-14B",
487
+ "scores": {
488
+ "average": 42.52,
489
+ "IFEval": 82.92,
490
+ "BBH": 49.75,
491
+ "MATH": 47.89,
492
+ "GPQA": 14.54,
493
+ "MUSR": 12.26,
494
+ "MMLU_PRO": 47.76,
495
+ "Architecture": "Qwen2ForCausalLM",
496
+ "Parameters": "14.766B",
497
+ "Chat_Template": "Yes"
498
+ },
499
+ "hf_url": "https://huggingface.co/djuna/Q2.5-Veltha-14B",
500
+ "known_config": null
501
+ },
502
+ {
503
+ "rank": 11,
504
+ "name": "arcee-ai/Virtuoso-Small-v2",
505
+ "scores": {
506
+ "average": 42.48,
507
+ "IFEval": 82.73,
508
+ "BBH": 50.95,
509
+ "MATH": 46.60,
510
+ "GPQA": 13.76,
511
+ "MUSR": 14.28,
512
+ "MMLU_PRO": 46.53,
513
+ "Architecture": "Qwen2ForCausalLM",
514
+ "Parameters": "14.766B",
515
+ "Chat_Template": "Yes"
516
+ },
517
+ "hf_url": "https://huggingface.co/arcee-ai/Virtuoso-Small-v2",
518
+ "known_config": null
519
+ },
520
+ {
521
+ "rank": 12,
522
+ "name": "YOYO-AI/Qwen2.5-14B-YOYO-V4-p1",
523
+ "scores": {
524
+ "average": 42.46,
525
+ "IFEval": 82.03,
526
+ "BBH": 50.25,
527
+ "MATH": 53.32,
528
+ "GPQA": 12.75,
529
+ "MUSR": 11.73,
530
+ "MMLU_PRO": 44.67,
531
+ "Architecture": "Qwen2ForCausalLM",
532
+ "Parameters": "14.766B",
533
+ "Chat_Template": "Yes"
534
+ },
535
+ "hf_url": "https://huggingface.co/YOYO-AI/Qwen2.5-14B-YOYO-V4-p1",
536
+ "known_config": null
537
+ },
538
+ {
539
+ "rank": 13,
540
+ "name": "jpacifico/Chocolatine-14B-Instruct-DPO-v1.3",
541
+ "scores": {
542
+ "average": 42.42,
543
+ "IFEval": 70.40,
544
+ "BBH": 54.85,
545
+ "MATH": 56.19,
546
+ "GPQA": 12.19,
547
+ "MUSR": 12.29,
548
+ "MMLU_PRO": 48.60,
549
+ "Architecture": "Phi3ForCausalLM",
550
+ "Parameters": "14.66B",
551
+ "Chat_Template": "Yes"
552
+ },
553
+ "hf_url": "https://huggingface.co/jpacifico/Chocolatine-14B-Instruct-DPO-v1.3",
554
+ "known_config": null
555
+ },
556
+ {
557
+ "rank": 14,
558
+ "name": "sthenno-com/miscii-14b-1028",
559
+ "scores": {
560
+ "average": 42.38,
561
+ "IFEval": 82.37,
562
+ "BBH": 49.26,
563
+ "MATH": 50.30,
564
+ "GPQA": 14.21,
565
+ "MUSR": 12.00,
566
+ "MMLU_PRO": 46.14,
567
+ "Architecture": "Qwen2ForCausalLM",
568
+ "Parameters": "14.77B",
569
+ "Chat_Template": "Yes"
570
+ },
571
+ "hf_url": "https://huggingface.co/sthenno-com/miscii-14b-1028",
572
+ "known_config": null
573
+ },
574
+ {
575
+ "rank": 15,
576
+ "name": "sthenno-com/miscii-14b-1225",
577
+ "scores": {
578
+ "average": 42.35,
579
+ "IFEval": 78.78,
580
+ "BBH": 50.91,
581
+ "MATH": 45.17,
582
+ "GPQA": 17.00,
583
+ "MUSR": 14.77,
584
+ "MMLU_PRO": 47.46,
585
+ "Architecture": "Qwen2ForCausalLM",
586
+ "Parameters": "14.766B",
587
+ "Chat_Template": "Yes"
588
+ },
589
+ "hf_url": "https://huggingface.co/sthenno-com/miscii-14b-1225",
590
+ "known_config": null
591
+ },
592
+ {
593
+ "rank": 16,
594
+ "name": "prithivMLmods/Sombrero-Opus-14B-Elite5",
595
+ "scores": {
596
+ "average": 42.32,
597
+ "IFEval": 78.81,
598
+ "BBH": 50.17,
599
+ "MATH": 53.55,
600
+ "GPQA": 11.52,
601
+ "MUSR": 13.22,
602
+ "MMLU_PRO": 46.67,
603
+ "Architecture": "Qwen2ForCausalLM",
604
+ "Parameters": "14.766B",
605
+ "Chat_Template": "Yes"
606
+ },
607
+ "hf_url": "https://huggingface.co/prithivMLmods/Sombrero-Opus-14B-Elite5",
608
+ "known_config": null
609
+ },
610
+ {
611
+ "rank": 17,
612
+ "name": "Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v8",
613
+ "scores": {
614
+ "average": 42.26,
615
+ "IFEval": 73.84,
616
+ "BBH": 49.31,
617
+ "MATH": 41.69,
618
+ "GPQA": 18.23,
619
+ "MUSR": 21.96,
620
+ "MMLU_PRO": 48.50,
621
+ "Architecture": "Qwen2ForCausalLM",
622
+ "Parameters": "14.766B",
623
+ "Chat_Template": "No"
624
+ },
625
+ "hf_url": "https://huggingface.co/Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v8",
626
+ "known_config": null
627
+ },
628
+ {
629
+ "rank": 18,
630
+ "name": "prithivMLmods/Equuleus-Opus-14B-Exp",
631
+ "scores": {
632
+ "average": 42.20,
633
+ "IFEval": 70.01,
634
+ "BBH": 48.62,
635
+ "MATH": 45.85,
636
+ "GPQA": 18.23,
637
+ "MUSR": 21.90,
638
+ "MMLU_PRO": 48.60,
639
+ "Architecture": "Qwen2ForCausalLM",
640
+ "Parameters": "14.766B",
641
+ "Chat_Template": "No"
642
+ },
643
+ "hf_url": "https://huggingface.co/prithivMLmods/Equuleus-Opus-14B-Exp",
644
+ "known_config": null
645
+ },
646
+ {
647
+ "rank": 19,
648
+ "name": "rombodawg/Rombos-LLM-V2.6-Qwen-14b",
649
+ "scores": {
650
+ "average": 42.20,
651
+ "IFEval": 84.32,
652
+ "BBH": 49.28,
653
+ "MATH": 52.11,
654
+ "GPQA": 11.19,
655
+ "MUSR": 12.29,
656
+ "MMLU_PRO": 44.01,
657
+ "Architecture": "Qwen2ForCausalLM",
658
+ "Parameters": "14.77B",
659
+ "Chat_Template": "Yes"
660
+ },
661
+ "hf_url": "https://huggingface.co/rombodawg/Rombos-LLM-V2.6-Qwen-14b",
662
+ "known_config": null
663
+ },
664
+ {
665
+ "rank": 20,
666
+ "name": "nbeerbower/EVA-abliterated-TIES-Qwen2.5-14B",
667
+ "scores": {
668
+ "average": 42.16,
669
+ "IFEval": 78.36,
670
+ "BBH": 48.52,
671
+ "MATH": 50.45,
672
+ "GPQA": 13.98,
673
+ "MUSR": 14.88,
674
+ "MMLU_PRO": 46.79,
675
+ "Architecture": "Qwen2ForCausalLM",
676
+ "Parameters": "14.77B",
677
+ "Chat_Template": "Yes"
678
+ },
679
+ "hf_url": "https://huggingface.co/nbeerbower/EVA-abliterated-TIES-Qwen2.5-14B",
680
+ "known_config": null
681
+ },
682
+ {
683
+ "rank": 21,
684
+ "name": "sometimesanotion/LamarckInfusion-14B-v1",
685
+ "scores": {
686
+ "average": 42.06,
687
+ "IFEval": 71.98,
688
+ "BBH": 50.35,
689
+ "MATH": 41.69,
690
+ "GPQA": 18.79,
691
+ "MUSR": 20.90,
692
+ "MMLU_PRO": 48.63,
693
+ "Architecture": "Qwen2ForCausalLM",
694
+ "Parameters": "14.766B",
695
+ "Chat_Template": "No"
696
+ },
697
+ "hf_url": "https://huggingface.co/sometimesanotion/LamarckInfusion-14B-v1",
698
+ "known_config": null
699
+ },
700
+ {
701
+ "rank": 22,
702
+ "name": "tensopolis/virtuoso-small-v2-tensopolis-v1",
703
+ "scores": {
704
+ "average": 41.99,
705
+ "IFEval": 82.40,
706
+ "BBH": 50.53,
707
+ "MATH": 46.53,
708
+ "GPQA": 12.53,
709
+ "MUSR": 13.88,
710
+ "MMLU_PRO": 46.07,
711
+ "Architecture": "Qwen2ForCausalLM",
712
+ "Parameters": "14.766B",
713
+ "Chat_Template": "Yes"
714
+ },
715
+ "hf_url": "https://huggingface.co/tensopolis/virtuoso-small-v2-tensopolis-v1",
716
+ "known_config": null
717
+ },
718
+ {
719
+ "rank": 23,
720
+ "name": "Quazim0t0/Fugazi14b",
721
+ "scores": {
722
+ "average": 41.94,
723
+ "IFEval": 69.98,
724
+ "BBH": 56.09,
725
+ "MATH": 46.53,
726
+ "GPQA": 13.53,
727
+ "MUSR": 16.42,
728
+ "MMLU_PRO": 49.08,
729
+ "Architecture": "LlamaForCausalLM",
730
+ "Parameters": "14.66B",
731
+ "Chat_Template": "Yes"
732
+ },
733
+ "hf_url": "https://huggingface.co/Quazim0t0/Fugazi14b",
734
+ "known_config": null
735
+ },
736
+ {
737
+ "rank": 24,
738
+ "name": "1024m/QWEN-14B-B100",
739
+ "scores": {
740
+ "average": 41.92,
741
+ "IFEval": 77.62,
742
+ "BBH": 49.78,
743
+ "MATH": 54.38,
744
+ "GPQA": 13.42,
745
+ "MUSR": 9.88,
746
+ "MMLU_PRO": 46.43,
747
+ "Architecture": "Qwen2ForCausalLM",
748
+ "Parameters": "14.77B",
749
+ "Chat_Template": "Yes"
750
+ },
751
+ "hf_url": "https://huggingface.co/1024m/QWEN-14B-B100",
752
+ "known_config": null
753
+ },
754
+ {
755
+ "rank": 25,
756
+ "name": "Sakalti/Saka-14B",
757
+ "scores": {
758
+ "average": 41.91,
759
+ "IFEval": 71.74,
760
+ "BBH": 49.72,
761
+ "MATH": 40.94,
762
+ "GPQA": 19.46,
763
+ "MUSR": 20.74,
764
+ "MMLU_PRO": 48.84,
765
+ "Architecture": "Qwen2ForCausalLM",
766
+ "Parameters": "14.766B",
767
+ "Chat_Template": "No"
768
+ },
769
+ "hf_url": "https://huggingface.co/Sakalti/Saka-14B",
770
+ "known_config": null
771
+ },
772
+ {
773
+ "rank": 26,
774
+ "name": "prithivMLmods/Sombrero-Opus-14B-Elite6",
775
+ "scores": {
776
+ "average": 41.88,
777
+ "IFEval": 72.26,
778
+ "BBH": 49.60,
779
+ "MATH": 40.79,
780
+ "GPQA": 19.13,
781
+ "MUSR": 20.74,
782
+ "MMLU_PRO": 48.78,
783
+ "Architecture": "Qwen2ForCausalLM",
784
+ "Parameters": "14.766B",
785
+ "Chat_Template": "No"
786
+ },
787
+ "hf_url": "https://huggingface.co/prithivMLmods/Sombrero-Opus-14B-Elite6",
788
+ "known_config": null
789
+ },
790
+ {
791
+ "rank": 27,
792
+ "name": "YOYO-AI/Qwen2.5-14B-YOYO-latest-V2",
793
+ "scores": {
794
+ "average": 41.85,
795
+ "IFEval": 77.71,
796
+ "BBH": 47.30,
797
+ "MATH": 51.59,
798
+ "GPQA": 13.87,
799
+ "MUSR": 13.68,
800
+ "MMLU_PRO": 46.93,
801
+ "Architecture": "Qwen2ForCausalLM",
802
+ "Parameters": "14.766B",
803
+ "Chat_Template": "Yes"
804
+ },
805
+ "hf_url": "https://huggingface.co/YOYO-AI/Qwen2.5-14B-YOYO-latest-V2",
806
+ "known_config": null
807
+ },
808
+ {
809
+ "rank": 28,
810
+ "name": "Tsunami-th/Tsunami-1.0-14B-Instruct",
811
+ "scores": {
812
+ "average": 41.84,
813
+ "IFEval": 78.29,
814
+ "BBH": 49.15,
815
+ "MATH": 45.85,
816
+ "GPQA": 14.21,
817
+ "MUSR": 16.34,
818
+ "MMLU_PRO": 47.21,
819
+ "Architecture": "Qwen2ForCausalLM",
820
+ "Parameters": "14.77B",
821
+ "Chat_Template": "Yes"
822
+ },
823
+ "hf_url": "https://huggingface.co/Tsunami-th/Tsunami-1.0-14B-Instruct",
824
+ "known_config": null
825
+ },
826
+ {
827
+ "rank": 29,
828
+ "name": "sthenno/tempesthenno-kto-0205-ckpt80",
829
+ "scores": {
830
+ "average": 41.79,
831
+ "IFEval": 80.54,
832
+ "BBH": 50.64,
833
+ "MATH": 45.92,
834
+ "GPQA": 13.09,
835
+ "MUSR": 12.93,
836
+ "MMLU_PRO": 47.62,
837
+ "Architecture": "Qwen2ForCausalLM",
838
+ "Parameters": "14.766B",
839
+ "Chat_Template": "No"
840
+ },
841
+ "hf_url": "https://huggingface.co/sthenno/tempesthenno-kto-0205-ckpt80",
842
+ "known_config": null
843
+ },
844
+ {
845
+ "rank": 30,
846
+ "name": "sometimesanotion/Lamarck-14B-v0.7-rc4",
847
+ "scores": {
848
+ "average": 41.79,
849
+ "IFEval": 72.11,
850
+ "BBH": 49.85,
851
+ "MATH": 40.26,
852
+ "GPQA": 18.57,
853
+ "MUSR": 21.07,
854
+ "MMLU_PRO": 48.89,
855
+ "Architecture": "Qwen2ForCausalLM",
856
+ "Parameters": "14.766B",
857
+ "Chat_Template": "No"
858
+ },
859
+ "hf_url": "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7-rc4",
860
+ "known_config": null
861
+ },
862
+ {
863
+ "rank": 31,
864
+ "name": "prithivMLmods/Porpoise-Opus-14B-Exp",
865
+ "scores": {
866
+ "average": 41.77,
867
+ "IFEval": 70.98,
868
+ "BBH": 49.95,
869
+ "MATH": 40.41,
870
+ "GPQA": 19.13,
871
+ "MUSR": 21.30,
872
+ "MMLU_PRO": 48.85,
873
+ "Architecture": "Qwen2ForCausalLM",
874
+ "Parameters": "14.766B",
875
+ "Chat_Template": "No"
876
+ },
877
+ "hf_url": "https://huggingface.co/prithivMLmods/Porpoise-Opus-14B-Exp",
878
+ "known_config": null
879
+ },
880
+ {
881
+ "rank": 32,
882
+ "name": "CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES",
883
+ "scores": {
884
+ "average": 41.77,
885
+ "IFEval": 82.40,
886
+ "BBH": 48.20,
887
+ "MATH": 53.17,
888
+ "GPQA": 9.96,
889
+ "MUSR": 12.65,
890
+ "MMLU_PRO": 44.21,
891
+ "Architecture": "Qwen2ForCausalLM",
892
+ "Parameters": "14.77B",
893
+ "Chat_Template": "Yes"
894
+ },
895
+ "hf_url": "https://huggingface.co/CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES",
896
+ "known_config": null
897
+ },
898
+ {
899
+ "rank": 33,
900
  "name": "suayptalha/Lamarckvergence-14B",
901
  "scores": {
902
  "average": 43.32,
 
911
  "Chat_Template": "Yes"
912
  },
913
  "hf_url": "https://huggingface.co/suayptalha/Lamarckvergence-14B",
914
+ "known_config": null
915
  },
916
  {
917
+ "rank": 34,
918
  "name": "sthenno/tempesthenno-ppo-ckpt40",
919
  "scores": {
920
  "average": 42.74,
 
929
  "Chat_Template": "Yes"
930
  },
931
  "hf_url": "https://huggingface.co/sthenno/tempesthenno-ppo-ckpt40",
932
+ "known_config": null
933
  },
934
  {
935
+ "rank": 35,
936
  "name": "tanliboy/lambda-qwen2.5-14b-dpo-test",
937
  "scores": {
938
  "average": 42.62,
 
947
  "Chat_Template": "Yes"
948
  },
949
  "hf_url": "https://huggingface.co/tanliboy/lambda-qwen2.5-14b-dpo-test",
950
+ "known_config": null
951
  },
952
  {
953
+ "rank": 36,
954
  "name": "sthenno/tempesthenno-nuslerp-001",
955
  "scores": {
956
  "average": 42.59,
 
965
  "Chat_Template": "Yes"
966
  },
967
  "hf_url": "https://huggingface.co/sthenno/tempesthenno-nuslerp-001",
968
+ "known_config": null
969
  },
970
  {
971
+ "rank": 37,
972
  "name": "Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4",
973
  "scores": {
974
  "average": 42.55,
 
983
  "Chat_Template": "Yes"
984
  },
985
  "hf_url": "https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4",
986
+ "known_config": null
987
  },
988
  {
989
+ "rank": 38,
990
  "name": "djuna/Q2.5-Veltha-14B",
991
  "scores": {
992
  "average": 42.52,
 
1001
  "Chat_Template": "Yes"
1002
  },
1003
  "hf_url": "https://huggingface.co/djuna/Q2.5-Veltha-14B",
1004
+ "known_config": null
1005
  },
1006
  {
1007
+ "rank": 39,
1008
  "name": "arcee-ai/Virtuoso-Small-v2",
1009
  "scores": {
1010
  "average": 42.48,
 
1019
  "Chat_Template": "Yes"
1020
  },
1021
  "hf_url": "https://huggingface.co/arcee-ai/Virtuoso-Small-v2",
1022
+ "known_config": null
1023
  },
1024
  {
1025
+ "rank": 40,
1026
  "name": "jpacifico/Chocolatine-14B-Instruct-DPO-v1.3",
1027
  "scores": {
1028
  "average": 42.42,
 
1037
  "Chat_Template": "Yes"
1038
  },
1039
  "hf_url": "https://huggingface.co/jpacifico/Chocolatine-14B-Instruct-DPO-v1.3",
1040
+ "known_config": null
1041
  },
1042
  {
1043
+ "rank": 41,
1044
  "name": "sthenno-com/miscii-14b-1028",
1045
  "scores": {
1046
  "average": 42.38,
 
1055
  "Chat_Template": "Yes"
1056
  },
1057
  "hf_url": "https://huggingface.co/sthenno-com/miscii-14b-1028",
1058
+ "known_config": null
1059
  },
1060
  {
1061
+ "rank": 42,
1062
  "name": "sthenno-com/miscii-14b-1225",
1063
  "scores": {
1064
  "average": 42.35,
 
1073
  "Chat_Template": "Yes"
1074
  },
1075
  "hf_url": "https://huggingface.co/sthenno-com/miscii-14b-1225",
1076
+ "known_config": null
1077
  },
1078
  {
1079
+ "rank": 43,
1080
  "name": "tensopolis/virtuoso-small-v2-tensopolis-v1",
1081
  "scores": {
1082
  "average": 42.34,
 
1091
  "Chat_Template": "Yes"
1092
  },
1093
  "hf_url": "https://huggingface.co/tensopolis/virtuoso-small-v2-tensopolis-v1",
1094
+ "known_config": null
1095
  },
1096
  {
1097
+ "rank": 44,
1098
  "name": "rombodawg/Rombos-LLM-V2.6-Qwen-14b",
1099
  "scores": {
1100
  "average": 42.2,
 
1109
  "Chat_Template": "Yes"
1110
  },
1111
  "hf_url": "https://huggingface.co/rombodawg/Rombos-LLM-V2.6-Qwen-14b",
1112
+ "known_config": null
1113
  },
1114
  {
1115
+ "rank": 45,
1116
  "name": "1024m/QWEN-14B-B100",
1117
  "scores": {
1118
  "average": 41.92,
 
1127
  "Chat_Template": "Yes"
1128
  },
1129
  "hf_url": "https://huggingface.co/1024m/QWEN-14B-B100",
1130
+ "known_config": null
1131
  },
1132
  {
1133
+ "rank": 46,
1134
  "name": "Sakalti/Saka-14B",
1135
  "scores": {
1136
  "average": 41.91,
 
1145
  "Chat_Template": "No"
1146
  },
1147
  "hf_url": "https://huggingface.co/Sakalti/Saka-14B",
1148
+ "known_config": null
1149
  },
1150
  {
1151
+ "rank": 47,
1152
  "name": "Tsunami-th/Tsunami-1.0-14B-Instruct",
1153
  "scores": {
1154
  "average": 41.84,
 
1163
  "Chat_Template": "Yes"
1164
  },
1165
  "hf_url": "https://huggingface.co/Tsunami-th/Tsunami-1.0-14B-Instruct",
1166
+ "known_config": null
1167
  },
1168
  {
1169
+ "rank": 48,
1170
  "name": "sthenno/tempesthenno-kto-0205-ckpt80",
1171
  "scores": {
1172
  "average": 41.79,
 
1181
  "Chat_Template": "No"
1182
  },
1183
  "hf_url": "https://huggingface.co/sthenno/tempesthenno-kto-0205-ckpt80",
1184
+ "known_config": null
1185
  },
1186
  {
1187
+ "rank": 49,
1188
  "name": "sometimesanotion/Lamarck-14B-v0.7-rc4",
1189
  "scores": {
1190
  "average": 41.79,
 
1199
  "Chat_Template": "No"
1200
  },
1201
  "hf_url": "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.7-rc4",
1202
+ "known_config": null
1203
  },
1204
  {
1205
+ "rank": 50,
1206
  "name": "CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES",
1207
  "scores": {
1208
  "average": 41.77,
 
1217
  "Chat_Template": "Yes"
1218
  },
1219
  "hf_url": "https://huggingface.co/CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES",
1220
+ "known_config": null
1221
  },
1222
  {
1223
+ "rank": 51,
1224
  "name": "suayptalha/Luminis-phi-4",
1225
  "scores": {
1226
  "average": 41.76,
 
1235
  "Chat_Template": "Yes"
1236
  },
1237
  "hf_url": "https://huggingface.co/suayptalha/Luminis-phi-4",
1238
+ "known_config": null
1239
  },
1240
  {
1241
+ "rank": 52,
1242
  "name": "huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2",
1243
  "scores": {
1244
  "average": 41.75,
 
1253
  "Chat_Template": "Yes"
1254
  },
1255
  "hf_url": "https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2",
1256
+ "known_config": null
1257
  },
1258
  {
1259
+ "rank": 53,
1260
  "name": "djuna/Q2.5-Veltha-14B-0.5",
1261
  "scores": {
1262
  "average": 41.61,
 
1271
  "Chat_Template": "Yes"
1272
  },
1273
  "hf_url": "https://huggingface.co/djuna/Q2.5-Veltha-14B-0.5",
1274
+ "known_config": null
1275
  },
1276
  {
1277
+ "rank": 54,
1278
  "name": "Qwen/Qwen2.5-14B-Instruct-1M",
1279
  "scores": {
1280
  "average": 41.56,
 
1289
  "Chat_Template": "Yes"
1290
  },
1291
  "hf_url": "https://huggingface.co/Qwen/Qwen2.5-14B-Instruct-1M",
1292
+ "known_config": null
1293
  },
1294
  {
1295
+ "rank": 55,
1296
  "name": "notbdq/Qwen2.5-14B-Instruct-1M-GRPO-Reasoning",
1297
  "scores": {
1298
  "average": 41.56,
 
1307
  "Chat_Template": "Yes"
1308
  },
1309
  "hf_url": "https://huggingface.co/notbdq/Qwen2.5-14B-Instruct-1M-GRPO-Reasoning",
1310
+ "known_config": null
1311
  },
1312
  {
1313
+ "rank": 56,
1314
  "name": "sometimesanotion/Qwenvergence-14B-v11",
1315
  "scores": {
1316
  "average": 41.52,
 
1325
  "Chat_Template": "No"
1326
  },
1327
  "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v11",
1328
+ "known_config": null
1329
  },
1330
  {
1331
+ "rank": 57,
1332
  "name": "sometimesanotion/Qwenvergence-14B-v10",
1333
  "scores": {
1334
  "average": 41.48,
 
1343
  "Chat_Template": "No"
1344
  },
1345
  "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v10",
1346
+ "known_config": null
1347
  },
1348
  {
1349
+ "rank": 58,
1350
  "name": "CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES",
1351
  "scores": {
1352
  "average": 41.47,
 
1361
  "Chat_Template": "Yes"
1362
  },
1363
  "hf_url": "https://huggingface.co/CombinHorizon/huihui-ai-abliteratedV2-Qwen2.5-14B-Inst-BaseMerge-TIES",
1364
+ "known_config": null
1365
  },
1366
  {
1367
+ "rank": 59,
1368
  "name": "RDson/WomboCombo-R1-Coder-14B-Preview",
1369
  "scores": {
1370
  "average": 41.46,
 
1379
  "Chat_Template": "Yes"
1380
  },
1381
  "hf_url": "https://huggingface.co/RDson/WomboCombo-R1-Coder-14B-Preview",
1382
+ "known_config": null
1383
  },
1384
  {
1385
+ "rank": 60,
1386
  "name": "jpacifico/Chocolatine-2-14B-Instruct-v2.0b3",
1387
  "scores": {
1388
  "average": 41.43,
 
1397
  "Chat_Template": "No"
1398
  },
1399
  "hf_url": "https://huggingface.co/jpacifico/Chocolatine-2-14B-Instruct-v2.0b3",
1400
+ "known_config": null
1401
  },
1402
  {
1403
+ "rank": 61,
1404
  "name": "Quazim0t0/Nova-14b-sce",
1405
  "scores": {
1406
  "average": 41.41,
 
1415
  "Chat_Template": "Yes"
1416
  },
1417
  "hf_url": "https://huggingface.co/Quazim0t0/Nova-14b-sce",
1418
+ "known_config": null
1419
  },
1420
  {
1421
+ "rank": 62,
1422
  "name": "v000000/Qwen2.5-14B-Gutenberg-Instruct-Slerpeno",
1423
  "scores": {
1424
  "average": 41.36,
 
1433
  "Chat_Template": "Yes"
1434
  },
1435
  "hf_url": "https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-Instruct-Slerpeno",
1436
+ "known_config": null
1437
  },
1438
  {
1439
+ "rank": 63,
1440
  "name": "Quazim0t0/NovaScotia-14b-stock",
1441
  "scores": {
1442
  "average": 41.35,
 
1451
  "Chat_Template": "Yes"
1452
  },
1453
  "hf_url": "https://huggingface.co/Quazim0t0/NovaScotia-14b-stock",
1454
+ "known_config": null
1455
  },
1456
  {
1457
+ "rank": 64,
1458
  "name": "Quazim0t0/ODB-14b-sce",
1459
  "scores": {
1460
  "average": 41.34,
 
1469
  "Chat_Template": "Yes"
1470
  },
1471
  "hf_url": "https://huggingface.co/Quazim0t0/ODB-14b-sce",
1472
+ "known_config": null
1473
  },
1474
  {
1475
+ "rank": 65,
1476
  "name": "LightningRodLabs/Flashlight-v1.1",
1477
  "scores": {
1478
  "average": 40.99,
 
1487
  "Chat_Template": "Yes"
1488
  },
1489
  "hf_url": "https://huggingface.co/LightningRodLabs/Flashlight-v1.1",
1490
+ "known_config": null
1491
  },
1492
  {
1493
+ "rank": 66,
1494
  "name": "Quazim0t0/Mithril-14B-sce",
1495
  "scores": {
1496
  "average": 40.98,
 
1505
  "Chat_Template": "Yes"
1506
  },
1507
  "hf_url": "https://huggingface.co/Quazim0t0/Mithril-14B-sce",
1508
+ "known_config": null
1509
  },
1510
  {
1511
+ "rank": 67,
1512
  "name": "Sakalti/ultiima-14B-v0.2",
1513
  "scores": {
1514
  "average": 40.96,
 
1523
  "Chat_Template": "No"
1524
  },
1525
  "hf_url": "https://huggingface.co/Sakalti/ultiima-14B-v0.2",
1526
+ "known_config": null
1527
  },
1528
  {
1529
+ "rank": 68,
1530
  "name": "bunnycore/Phi-4-ReasoningRP",
1531
  "scores": {
1532
  "average": 40.95,
 
1541
  "Chat_Template": "Yes"
1542
  },
1543
  "hf_url": "https://huggingface.co/bunnycore/Phi-4-ReasoningRP",
1544
+ "known_config": null
1545
  },
1546
  {
1547
+ "rank": 69,
1548
  "name": "dwikitheduck/gen-inst-1",
1549
  "scores": {
1550
  "average": 40.88,
 
1559
  "Chat_Template": "Yes"
1560
  },
1561
  "hf_url": "https://huggingface.co/dwikitheduck/gen-inst-1",
1562
+ "known_config": null
1563
  },
1564
  {
1565
+ "rank": 70,
1566
  "name": "v000000/Qwen2.5-14B-Gutenberg-1e-Delta",
1567
  "scores": {
1568
  "average": 40.88,
 
1577
  "Chat_Template": "Yes"
1578
  },
1579
  "hf_url": "https://huggingface.co/v000000/Qwen2.5-14B-Gutenberg-1e-Delta",
1580
+ "known_config": null
1581
  },
1582
  {
1583
  "rank": 60,
 
1595
  "Chat_Template": "No"
1596
  },
1597
  "hf_url": "https://huggingface.co/hotmailuser/QwenSlerp2-14B",
1598
+ "known_config": null
1599
  },
1600
  {
1601
+ "rank": 71,
1602
  "name": "Quazim0t0/Loke-14B-sce",
1603
  "scores": {
1604
  "average": 40.86,
 
1613
  "Chat_Template": "Yes"
1614
  },
1615
  "hf_url": "https://huggingface.co/Quazim0t0/Loke-14B-sce",
1616
+ "known_config": null
1617
  },
1618
  {
1619
+ "rank": 72,
1620
  "name": "Quazim0t0/mosaic-14b-sce",
1621
  "scores": {
1622
  "average": 40.83,
 
1631
  "Chat_Template": "Yes"
1632
  },
1633
  "hf_url": "https://huggingface.co/Quazim0t0/mosaic-14b-sce",
1634
+ "known_config": null
1635
  },
1636
  {
1637
+ "rank": 73,
1638
  "name": "bunnycore/Phi-4-Model-Stock",
1639
  "scores": {
1640
  "average": 40.79,
 
1649
  "Chat_Template": "Yes"
1650
  },
1651
  "hf_url": "https://huggingface.co/bunnycore/Phi-4-Model-Stock",
1652
+ "known_config": null
1653
  },
1654
  {
1655
+ "rank": 74,
1656
  "name": "unsloth/phi-4",
1657
  "scores": {
1658
  "average": 40.73,
 
1667
  "Chat_Template": "Yes"
1668
  },
1669
  "hf_url": "https://huggingface.co/unsloth/phi-4",
1670
+ "known_config": null
1671
  },
1672
  {
1673
+ "rank": 75,
1674
  "name": "pankajmathur/orca_mini_phi-4",
1675
  "scores": {
1676
  "average": 40.68,
 
1685
  "Chat_Template": "Yes"
1686
  },
1687
  "hf_url": "https://huggingface.co/pankajmathur/orca_mini_phi-4",
1688
+ "known_config": null
1689
  },
1690
  {
1691
+ "rank": 76,
1692
  "name": "pankajmathur/orca_mini_v9_2_14B",
1693
  "scores": {
1694
  "average": 40.68,
 
1703
  "Chat_Template": "Yes"
1704
  },
1705
  "hf_url": "https://huggingface.co/pankajmathur/orca_mini_v9_2_14B",
1706
+ "known_config": null
1707
  },
1708
  {
1709
+ "rank": 77,
1710
  "name": "sometimesanotion/Lamarck-14B-v0.6-model_stock",
1711
  "scores": {
1712
  "average": 40.68,
 
1721
  "Chat_Template": "No"
1722
  },
1723
  "hf_url": "https://huggingface.co/sometimesanotion/Lamarck-14B-v0.6-model_stock",
1724
+ "known_config": null
1725
  },
1726
  {
1727
+ "rank": 78,
1728
  "name": "sometimesanotion/Qwenvergence-14B-v0.6-004-model_stock",
1729
  "scores": {
1730
  "average": 40.6,
 
1739
  "Chat_Template": "No"
1740
  },
1741
  "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-v0.6-004-model_stock",
1742
+ "known_config": null
1743
  },
1744
  {
1745
+ "rank": 79,
1746
  "name": "Quazim0t0/Oasis-14B-ties",
1747
  "scores": {
1748
  "average": 40.59,
 
1757
  "Chat_Template": "Yes"
1758
  },
1759
  "hf_url": "https://huggingface.co/Quazim0t0/Oasis-14B-ties",
1760
+ "known_config": null
1761
  },
1762
  {
1763
+ "rank": 80,
1764
  "name": "LightningRodLabs/Flashlight-v1.0",
1765
  "scores": {
1766
  "average": 40.57,
 
1775
  "Chat_Template": "Yes"
1776
  },
1777
  "hf_url": "https://huggingface.co/LightningRodLabs/Flashlight-v1.0",
1778
+ "known_config": null
1779
  },
1780
  {
1781
+ "rank": 81,
1782
  "name": "arcee-ai/Virtuoso-Small",
1783
  "scores": {
1784
  "average": 40.54,
 
1793
  "Chat_Template": "Yes"
1794
  },
1795
  "hf_url": "https://huggingface.co/arcee-ai/Virtuoso-Small",
1796
+ "known_config": null
1797
  },
1798
  {
1799
+ "rank": 82,
1800
  "name": "Quazim0t0/GuiltySpark-14B-ties",
1801
  "scores": {
1802
  "average": 40.52,
 
1811
  "Chat_Template": "Yes"
1812
  },
1813
  "hf_url": "https://huggingface.co/Quazim0t0/GuiltySpark-14B-ties",
1814
+ "known_config": null
1815
  },
1816
  {
1817
+ "rank": 83,
1818
  "name": "ozone-ai/0x-lite",
1819
  "scores": {
1820
  "average": 40.48,
 
1829
  "Chat_Template": "Yes"
1830
  },
1831
  "hf_url": "https://huggingface.co/ozone-ai/0x-lite",
1832
+ "known_config": null
1833
  },
1834
  {
1835
+ "rank": 84,
1836
  "name": "Quazim0t0/Casa-14b-sce",
1837
  "scores": {
1838
  "average": 40.41,
 
1847
  "Chat_Template": "Yes"
1848
  },
1849
  "hf_url": "https://huggingface.co/Quazim0t0/Casa-14b-sce",
1850
+ "known_config": null
1851
  },
1852
  {
1853
+ "rank": 85,
1854
  "name": "Sakalti/ultiima-14B-v0.3",
1855
  "scores": {
1856
  "average": 40.38,
 
1865
  "Chat_Template": "No"
1866
  },
1867
  "hf_url": "https://huggingface.co/Sakalti/ultiima-14B-v0.3",
1868
+ "known_config": null
1869
  },
1870
  {
1871
+ "rank": 86,
1872
  "name": "ehristoforu/fp4-14b-v1-fix",
1873
  "scores": {
1874
  "average": 40.37,
 
1883
  "Chat_Template": "Yes"
1884
  },
1885
  "hf_url": "https://huggingface.co/ehristoforu/fp4-14b-v1-fix",
1886
+ "known_config": null
1887
  },
1888
  {
1889
+ "rank": 87,
1890
  "name": "FINGU-AI/Chocolatine-Fusion-14B",
1891
  "scores": {
1892
  "average": 40.36,
 
1901
  "Chat_Template": "No"
1902
  },
1903
  "hf_url": "https://huggingface.co/FINGU-AI/Chocolatine-Fusion-14B",
1904
+ "known_config": null
1905
  },
1906
  {
1907
+ "rank": 88,
1908
  "name": "hotmailuser/QwenSlerp-14B",
1909
  "scores": {
1910
  "average": 40.35,
 
1919
  "Chat_Template": "No"
1920
  },
1921
  "hf_url": "https://huggingface.co/hotmailuser/QwenSlerp-14B",
1922
+ "known_config": null
1923
  },
1924
  {
1925
+ "rank": 89,
1926
  "name": "Triangle104/Robo-Gutenberg_V1.0",
1927
  "scores": {
1928
  "average": 40.35,
 
1937
  "Chat_Template": "No"
1938
  },
1939
  "hf_url": "https://huggingface.co/Triangle104/Robo-Gutenberg_V1.0",
1940
+ "known_config": null
1941
  },
1942
  {
1943
+ "rank": 90,
1944
  "name": "Quazim0t0/Adamant-14B-sce",
1945
  "scores": {
1946
  "average": 40.32,
 
1955
  "Chat_Template": "Yes"
1956
  },
1957
  "hf_url": "https://huggingface.co/Quazim0t0/Adamant-14B-sce",
1958
+ "known_config": null
1959
  },
1960
  {
1961
+ "rank": 91,
1962
  "name": "Quazim0t0/Phi4Basis-14B-sce",
1963
  "scores": {
1964
  "average": 40.31,
 
1973
  "Chat_Template": "Yes"
1974
  },
1975
  "hf_url": "https://huggingface.co/Quazim0t0/Phi4Basis-14B-sce",
1976
+ "known_config": null
1977
  },
1978
  {
1979
+ "rank": 92,
1980
  "name": "Quazim0t0/bloom-14b-stock",
1981
  "scores": {
1982
  "average": 40.29,
 
1991
  "Chat_Template": "Yes"
1992
  },
1993
  "hf_url": "https://huggingface.co/Quazim0t0/bloom-14b-stock",
1994
+ "known_config": null
1995
  },
1996
  {
1997
+ "rank": 93,
1998
  "name": "sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-Prose01",
1999
  "scores": {
2000
  "average": 40.28,
 
2009
  "Chat_Template": "No"
2010
  },
2011
  "hf_url": "https://huggingface.co/sometimesanotion/Qwen2.5-14B-Vimarckoso-v3-Prose01",
2012
+ "known_config": null
2013
  },
2014
  {
2015
+ "rank": 94,
2016
  "name": "Quazim0t0/Halo-14B-sce",
2017
  "scores": {
2018
  "average": 40.26,
 
2027
  "Chat_Template": "Yes"
2028
  },
2029
  "hf_url": "https://huggingface.co/Quazim0t0/Halo-14B-sce",
2030
+ "known_config": null
2031
  },
2032
  {
2033
+ "rank": 95,
2034
  "name": "prithivMLmods/Calcium-Opus-14B-Elite2",
2035
  "scores": {
2036
  "average": 40.25,
 
2045
  "Chat_Template": "No"
2046
  },
2047
  "hf_url": "https://huggingface.co/prithivMLmods/Calcium-Opus-14B-Elite2",
2048
+ "known_config": null
2049
  },
2050
  {
2051
+ "rank": 96,
2052
  "name": "SicariusSicariiStuff/Impish_QWEN_14B-1M",
2053
  "scores": {
2054
  "average": 40.24,
 
2063
  "Chat_Template": "Yes"
2064
  },
2065
  "hf_url": "https://huggingface.co/SicariusSicariiStuff/Impish_QWEN_14B-1M",
2066
+ "known_config": null
2067
  },
2068
  {
2069
+ "rank": 97,
2070
  "name": "bunnycore/Phi-4-Stock-Ex",
2071
  "scores": {
2072
  "average": 40.22,
 
2081
  "Chat_Template": "Yes"
2082
  },
2083
  "hf_url": "https://huggingface.co/bunnycore/Phi-4-Stock-Ex",
2084
+ "known_config": null
2085
  },
2086
  {
2087
+ "rank": 98,
2088
  "name": "sometimesanotion/Qwenvergence-14B-qv256",
2089
  "scores": {
2090
  "average": 40.12,
 
2099
  "Chat_Template": "No"
2100
  },
2101
  "hf_url": "https://huggingface.co/sometimesanotion/Qwenvergence-14B-qv256",
2102
+ "known_config": null
2103
  },
2104
  {
2105
+ "rank": 99,
2106
  "name": "tensopolis/virtuoso-small-tensopolis-v2",
2107
  "scores": {
2108
  "average": 40.11,
 
2117
  "Chat_Template": "Yes"
2118
  },
2119
  "hf_url": "https://huggingface.co/tensopolis/virtuoso-small-tensopolis-v2",
2120
+ "known_config": null
2121
  }
2122
  ]
2123
 
2124
 
2125
 
2126
+
2127
  def snippet_scrape_model_page(url):
2128
  """
2129
  Equivalent scraping function for the larger dataset