CultriX commited on
Commit
6acb805
·
verified ·
1 Parent(s): 06fd517

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +512 -0
app.py CHANGED
@@ -319,6 +319,518 @@ def download_all_data():
319
  # --------------------------------------------------------------------
320
  # This is your larger dataset, rank = 44..105
321
  benchmark_data = [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
322
  {
323
  "rank": 1,
324
  "name": "suayptalha/Lamarckvergence-14B",
 
319
  # --------------------------------------------------------------------
320
  # This is your larger dataset, rank = 44..105
321
  benchmark_data = [
322
+ {
323
+ "Model Rank": "1",
324
+ "Icon": "ð¤",
325
+ "Model Name": "wanlige/li-14b-v0.4",
326
+ "Average Score": "43.66 %",
327
+ "IFEval Score": "81.33 %",
328
+ "BBH Score": "50.38 %",
329
+ "MATH Score": "55.74 %",
330
+ "GPQA Score": "11.86 %",
331
+ "MUSR Score": "16.35 %",
332
+ "MMLU-PRO Score": "46.30 %",
333
+ "Model Architecture": "Qwen2ForCausalLM",
334
+ "Model Precision": "bfloat16",
335
+ "Model Parameters": "14.77B",
336
+ "Chat Template Use": "Yes"
337
+ },
338
+ {
339
+ "Model Rank": "2",
340
+ "Icon": "ð¤",
341
+ "Model Name": "suayptalha/Lamarckvergence-14B",
342
+ "Average Score": "43.32 %",
343
+ "IFEval Score": "76.56 %",
344
+ "BBH Score": "50.33 %",
345
+ "MATH Score": "54.00 %",
346
+ "GPQA Score": "15.10 %",
347
+ "MUSR Score": "16.34 %",
348
+ "MMLU-PRO Score": "47.59 %",
349
+ "Model Architecture": "Qwen2ForCausalLM",
350
+ "Model Precision": "bfloat16",
351
+ "Model Parameters": "14.766B",
352
+ "Chat Template Use": "Yes"
353
+ },
354
+ {
355
+ "Model Rank": "3",
356
+ "Icon": "ð¤",
357
+ "Model Name": "wanlige/li-14b-v0.4-slerp0.1",
358
+ "Average Score": "42.91 %",
359
+ "IFEval Score": "79.23 %",
360
+ "BBH Score": "50.88 %",
361
+ "MATH Score": "53.32 %",
362
+ "GPQA Score": "14.54 %",
363
+ "MUSR Score": "11.75 %",
364
+ "MMLU-PRO Score": "47.71 %",
365
+ "Model Architecture": "Qwen2ForCausalLM",
366
+ "Model Precision": "bfloat16",
367
+ "Model Parameters": "14.766B",
368
+ "Chat Template Use": "Yes"
369
+ },
370
+ {
371
+ "Model Rank": "4",
372
+ "Icon": "ð¬",
373
+ "Model Name": "sthenno-com/miscii-14b-0218",
374
+ "Average Score": "42.90 %",
375
+ "IFEval Score": "76.56 %",
376
+ "BBH Score": "50.64 %",
377
+ "MATH Score": "51.44 %",
378
+ "GPQA Score": "17.79 %",
379
+ "MUSR Score": "13.21 %",
380
+ "MMLU-PRO Score": "47.75 %",
381
+ "Model Architecture": "Qwen2ForCausalLM",
382
+ "Model Precision": "bfloat16",
383
+ "Model Parameters": "14.766B",
384
+ "Chat Template Use": "Yes"
385
+ },
386
+ {
387
+ "Model Rank": "5",
388
+ "Icon": "├░┬╢",
389
+ "Model Name": "sthenno/tempesthenno-ppo-ckpt40",
390
+ "Average Score": "42.74 %",
391
+ "IFEval Score": "79.23 %",
392
+ "BBH Score": "50.57 %",
393
+ "MATH Score": "47.36 %",
394
+ "GPQA Score": "17.00 %",
395
+ "MUSR Score": "14.56 %",
396
+ "MMLU-PRO Score": "47.69 %",
397
+ "Model Architecture": "Qwen2ForCausalLM",
398
+ "Model Precision": "bfloat16",
399
+ "Model Parameters": "14.766B",
400
+ "Chat Template Use": "Yes"
401
+ },
402
+ {
403
+ "Model Rank": "6",
404
+ "Icon": "ð¬",
405
+ "Model Name": "tanliboy/lambda-qwen2.5-14b-dpo-test",
406
+ "Average Score": "42.62 %",
407
+ "IFEval Score": "82.31 %",
408
+ "BBH Score": "48.45 %",
409
+ "MATH Score": "54.61 %",
410
+ "GPQA Score": "14.99 %",
411
+ "MUSR Score": "12.59 %",
412
+ "MMLU-PRO Score": "42.75 %",
413
+ "Model Architecture": "Qwen2ForCausalLM",
414
+ "Model Precision": "bfloat16",
415
+ "Model Parameters": "14.77B",
416
+ "Chat Template Use": "Yes"
417
+ },
418
+ {
419
+ "Model Rank": "7",
420
+ "Icon": "├░┬╢",
421
+ "Model Name": "sthenno/tempesthenno-nuslerp-001",
422
+ "Average Score": "42.59 %",
423
+ "IFEval Score": "79.26 %",
424
+ "BBH Score": "51.04 %",
425
+ "MATH Score": "47.58 %",
426
+ "GPQA Score": "16.44 %",
427
+ "MUSR Score": "13.88 %",
428
+ "MMLU-PRO Score": "47.30 %",
429
+ "Model Architecture": "Qwen2ForCausalLM",
430
+ "Model Precision": "bfloat16",
431
+ "Model Parameters": "14.766B",
432
+ "Chat Template Use": "Yes"
433
+ },
434
+ {
435
+ "Model Rank": "8",
436
+ "Icon": "ð¤",
437
+ "Model Name": "YOYO-AI/Qwen2.5-14B-1M-YOYO-V3",
438
+ "Average Score": "42.56 %",
439
+ "IFEval Score": "83.98 %",
440
+ "BBH Score": "49.47 %",
441
+ "MATH Score": "53.55 %",
442
+ "GPQA Score": "10.51 %",
443
+ "MUSR Score": "11.10 %",
444
+ "MMLU-PRO Score": "46.74 %",
445
+ "Model Architecture": "Qwen2ForCausalLM",
446
+ "Model Precision": "float16",
447
+ "Model Parameters": "14.766B",
448
+ "Chat Template Use": "Yes"
449
+ },
450
+ {
451
+ "Model Rank": "9",
452
+ "Icon": "├░┬╢",
453
+ "Model Name": "Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4",
454
+ "Average Score": "42.55 %",
455
+ "IFEval Score": "82.92 %",
456
+ "BBH Score": "48.05 %",
457
+ "MATH Score": "54.23 %",
458
+ "GPQA Score": "12.30 %",
459
+ "MUSR Score": "13.15 %",
460
+ "MMLU-PRO Score": "44.65 %",
461
+ "Model Architecture": "Qwen2ForCausalLM",
462
+ "Model Precision": "bfloat16",
463
+ "Model Parameters": "14.77B",
464
+ "Chat Template Use": "Yes"
465
+ },
466
+ {
467
+ "Model Rank": "10",
468
+ "Icon": "ð¤",
469
+ "Model Name": "djuna/Q2.5-Veltha-14B",
470
+ "Average Score": "42.52 %",
471
+ "IFEval Score": "82.92 %",
472
+ "BBH Score": "49.75 %",
473
+ "MATH Score": "47.89 %",
474
+ "GPQA Score": "14.54 %",
475
+ "MUSR Score": "12.26 %",
476
+ "MMLU-PRO Score": "47.76 %",
477
+ "Model Architecture": "Qwen2ForCausalLM",
478
+ "Model Precision": "bfloat16",
479
+ "Model Parameters": "14.766B",
480
+ "Chat Template Use": "Yes"
481
+ },
482
+ {
483
+ "Model Rank": "11",
484
+ "Icon": "├░┬╢",
485
+ "Model Name": "arcee-ai/Virtuoso-Small-v2",
486
+ "Average Score": "42.48 %",
487
+ "IFEval Score": "82.73 %",
488
+ "BBH Score": "50.95 %",
489
+ "MATH Score": "46.60 %",
490
+ "GPQA Score": "13.76 %",
491
+ "MUSR Score": "14.28 %",
492
+ "MMLU-PRO Score": "46.53 %",
493
+ "Model Architecture": "Qwen2ForCausalLM",
494
+ "Model Precision": "bfloat16",
495
+ "Model Parameters": "14.766B",
496
+ "Chat Template Use": "Yes"
497
+ },
498
+ {
499
+ "Model Rank": "12",
500
+ "Icon": "ð¤",
501
+ "Model Name": "YOYO-AI/Qwen2.5-14B-YOYO-V4-p1",
502
+ "Average Score": "42.46 %",
503
+ "IFEval Score": "82.03 %",
504
+ "BBH Score": "50.25 %",
505
+ "MATH Score": "53.32 %",
506
+ "GPQA Score": "12.75 %",
507
+ "MUSR Score": "11.73 %",
508
+ "MMLU-PRO Score": "44.67 %",
509
+ "Model Architecture": "Qwen2ForCausalLM",
510
+ "Model Precision": "float16",
511
+ "Model Parameters": "14.766B",
512
+ "Chat Template Use": "Yes"
513
+ },
514
+ {
515
+ "Model Rank": "13",
516
+ "Icon": "ð¬",
517
+ "Model Name": "jpacifico/Chocolatine-14B-Instruct-DPO-v1.3",
518
+ "Average Score": "42.42 %",
519
+ "IFEval Score": "70.40 %",
520
+ "BBH Score": "54.85 %",
521
+ "MATH Score": "56.19 %",
522
+ "GPQA Score": "12.19 %",
523
+ "MUSR Score": "12.29 %",
524
+ "MMLU-PRO Score": "48.60 %",
525
+ "Model Architecture": "Phi3ForCausalLM",
526
+ "Model Precision": "float16",
527
+ "Model Parameters": "14.66B",
528
+ "Chat Template Use": "Yes"
529
+ },
530
+ {
531
+ "Model Rank": "14",
532
+ "Icon": "ð¬",
533
+ "Model Name": "sthenno-com/miscii-14b-1028",
534
+ "Average Score": "42.38 %",
535
+ "IFEval Score": "82.37 %",
536
+ "BBH Score": "49.26 %",
537
+ "MATH Score": "50.30 %",
538
+ "GPQA Score": "14.21 %",
539
+ "MUSR Score": "12.00 %",
540
+ "MMLU-PRO Score": "46.14 %",
541
+ "Model Architecture": "Qwen2ForCausalLM",
542
+ "Model Precision": "bfloat16",
543
+ "Model Parameters": "14.77B",
544
+ "Chat Template Use": "Yes"
545
+ },
546
+ {
547
+ "Model Rank": "15",
548
+ "Icon": "ð¬",
549
+ "Model Name": "sthenno-com/miscii-14b-1225",
550
+ "Average Score": "42.35 %",
551
+ "IFEval Score": "78.78 %",
552
+ "BBH Score": "50.91 %",
553
+ "MATH Score": "45.17 %",
554
+ "GPQA Score": "17.00 %",
555
+ "MUSR Score": "14.77 %",
556
+ "MMLU-PRO Score": "47.46 %",
557
+ "Model Architecture": "Qwen2ForCausalLM",
558
+ "Model Precision": "bfloat16",
559
+ "Model Parameters": "14.766B",
560
+ "Chat Template Use": "Yes"
561
+ },
562
+ {
563
+ "Model Rank": "16",
564
+ "Icon": "├░┬╢",
565
+ "Model Name": "prithivMLmods/Sombrero-Opus-14B-Elite5",
566
+ "Average Score": "42.32 %",
567
+ "IFEval Score": "78.81 %",
568
+ "BBH Score": "50.17 %",
569
+ "MATH Score": "53.55 %",
570
+ "GPQA Score": "11.52 %",
571
+ "MUSR Score": "13.22 %",
572
+ "MMLU-PRO Score": "46.67 %",
573
+ "Model Architecture": "Qwen2ForCausalLM",
574
+ "Model Precision": "float16",
575
+ "Model Parameters": "14.766B",
576
+ "Chat Template Use": "Yes"
577
+ },
578
+ {
579
+ "Model Rank": "17",
580
+ "Icon": "ð¤",
581
+ "Model Name": "Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v8",
582
+ "Average Score": "42.26 %",
583
+ "IFEval Score": "73.84 %",
584
+ "BBH Score": "49.31 %",
585
+ "MATH Score": "41.69 %",
586
+ "GPQA Score": "18.23 %",
587
+ "MUSR Score": "21.96 %",
588
+ "MMLU-PRO Score": "48.50 %",
589
+ "Model Architecture": "Qwen2ForCausalLM",
590
+ "Model Precision": "bfloat16",
591
+ "Model Parameters": "14.766B",
592
+ "Chat Template Use": "No"
593
+ },
594
+ {
595
+ "Model Rank": "18",
596
+ "Icon": "├░┬╢",
597
+ "Model Name": "prithivMLmods/Equuleus-Opus-14B-Exp",
598
+ "Average Score": "42.20 %",
599
+ "IFEval Score": "70.01 %",
600
+ "BBH Score": "48.62 %",
601
+ "MATH Score": "45.85 %",
602
+ "GPQA Score": "18.23 %",
603
+ "MUSR Score": "21.90 %",
604
+ "MMLU-PRO Score": "48.60 %",
605
+ "Model Architecture": "Qwen2ForCausalLM",
606
+ "Model Precision": "bfloat16",
607
+ "Model Parameters": "14.766B",
608
+ "Chat Template Use": "No"
609
+ },
610
+ {
611
+ "Model Rank": "19",
612
+ "Icon": "├░┬╢",
613
+ "Model Name": "rombodawg/Rombos-LLM-V2.6-Qwen-14b",
614
+ "Average Score": "42.20 %",
615
+ "IFEval Score": "84.32 %",
616
+ "BBH Score": "49.28 %",
617
+ "MATH Score": "52.11 %",
618
+ "GPQA Score": "11.19 %",
619
+ "MUSR Score": "12.29 %",
620
+ "MMLU-PRO Score": "44.01 %",
621
+ "Model Architecture": "Qwen2ForCausalLM",
622
+ "Model Precision": "bfloat16",
623
+ "Model Parameters": "14.77B",
624
+ "Chat Template Use": "Yes"
625
+ },
626
+ {
627
+ "Model Rank": "20",
628
+ "Icon": "ð¤",
629
+ "Model Name": "nbeerbower/EVA-abliterated-TIES-Qwen2.5-14B",
630
+ "Average Score": "42.16 %",
631
+ "IFEval Score": "78.36 %",
632
+ "BBH Score": "48.52 %",
633
+ "MATH Score": "50.45 %",
634
+ "GPQA Score": "13.98 %",
635
+ "MUSR Score": "14.88 %",
636
+ "MMLU-PRO Score": "46.79 %",
637
+ "Model Architecture": "Qwen2ForCausalLM",
638
+ "Model Precision": "bfloat16",
639
+ "Model Parameters": "14.77B",
640
+ "Chat Template Use": "Yes"
641
+ },
642
+ {
643
+ "Model Rank": "21",
644
+ "Icon": "ð¤",
645
+ "Model Name": "sometimesanotion/LamarckInfusion-14B-v1",
646
+ "Average Score": "42.06 %",
647
+ "IFEval Score": "71.98 %",
648
+ "BBH Score": "50.35 %",
649
+ "MATH Score": "41.69 %",
650
+ "GPQA Score": "18.79 %",
651
+ "MUSR Score": "20.90 %",
652
+ "MMLU-PRO Score": "48.63 %",
653
+ "Model Architecture": "Qwen2ForCausalLM",
654
+ "Model Precision": "bfloat16",
655
+ "Model Parameters": "14.766B",
656
+ "Chat Template Use": "No"
657
+ },
658
+ {
659
+ "Model Rank": "22",
660
+ "Icon": "ð¤",
661
+ "Model Name": "tensopolis/virtuoso-small-v2-tensopolis-v1",
662
+ "Average Score": "41.99 %",
663
+ "IFEval Score": "82.40 %",
664
+ "BBH Score": "50.53 %",
665
+ "MATH Score": "46.53 %",
666
+ "GPQA Score": "12.53 %",
667
+ "MUSR Score": "13.88 %",
668
+ "MMLU-PRO Score": "46.07 %",
669
+ "Model Architecture": "Qwen2ForCausalLM",
670
+ "Model Precision": "bfloat16",
671
+ "Model Parameters": "14.766B",
672
+ "Chat Template Use": "Yes"
673
+ },
674
+ {
675
+ "Model Rank": "23",
676
+ "Icon": "ð¤",
677
+ "Model Name": "Quazim0t0/Fugazi14b",
678
+ "Average Score": "41.94 %",
679
+ "IFEval Score": "69.98 %",
680
+ "BBH Score": "56.09 %",
681
+ "MATH Score": "46.53 %",
682
+ "GPQA Score": "13.53 %",
683
+ "MUSR Score": "16.42 %",
684
+ "MMLU-PRO Score": "49.08 %",
685
+ "Model Architecture": "LlamaForCausalLM",
686
+ "Model Precision": "bfloat16",
687
+ "Model Parameters": "14.66B",
688
+ "Chat Template Use": "Yes"
689
+ },
690
+ {
691
+ "Model Rank": "24",
692
+ "Icon": "├░┬╢",
693
+ "Model Name": "1024m/QWEN-14B-B100",
694
+ "Average Score": "41.92 %",
695
+ "IFEval Score": "77.62 %",
696
+ "BBH Score": "49.78 %",
697
+ "MATH Score": "54.38 %",
698
+ "GPQA Score": "13.42 %",
699
+ "MUSR Score": "9.88 %",
700
+ "MMLU-PRO Score": "46.43 %",
701
+ "Model Architecture": "Qwen2ForCausalLM",
702
+ "Model Precision": "bfloat16",
703
+ "Model Parameters": "14.77B",
704
+ "Chat Template Use": "Yes"
705
+ },
706
+ {
707
+ "Model Rank": "25",
708
+ "Icon": "├░┬╢",
709
+ "Model Name": "Sakalti/Saka-14B",
710
+ "Average Score": "41.91 %",
711
+ "IFEval Score": "71.74 %",
712
+ "BBH Score": "49.72 %",
713
+ "MATH Score": "40.94 %",
714
+ "GPQA Score": "19.46 %",
715
+ "MUSR Score": "20.74 %",
716
+ "MMLU-PRO Score": "48.84 %",
717
+ "Model Architecture": "Qwen2ForCausalLM",
718
+ "Model Precision": "float16",
719
+ "Model Parameters": "14.766B",
720
+ "Chat Template Use": "No"
721
+ },
722
+ {
723
+ "Model Rank": "26",
724
+ "Icon": "├░┬╢",
725
+ "Model Name": "prithivMLmods/Sombrero-Opus-14B-Elite6",
726
+ "Average Score": "41.88 %",
727
+ "IFEval Score": "72.26 %",
728
+ "BBH Score": "49.60 %",
729
+ "MATH Score": "40.79 %",
730
+ "GPQA Score": "19.13 %",
731
+ "MUSR Score": "20.74 %",
732
+ "MMLU-PRO Score": "48.78 %",
733
+ "Model Architecture": "Qwen2ForCausalLM",
734
+ "Model Precision": "bfloat16",
735
+ "Model Parameters": "14.766B",
736
+ "Chat Template Use": "No"
737
+ },
738
+ {
739
+ "Model Rank": "27",
740
+ "Icon": "ð¤",
741
+ "Model Name": "YOYO-AI/Qwen2.5-14B-YOYO-latest-V2",
742
+ "Average Score": "41.85 %",
743
+ "IFEval Score": "77.71 %",
744
+ "BBH Score": "47.30 %",
745
+ "MATH Score": "51.59 %",
746
+ "GPQA Score": "13.87 %",
747
+ "MUSR Score": "13.68 %",
748
+ "MMLU-PRO Score": "46.93 %",
749
+ "Model Architecture": "Qwen2ForCausalLM",
750
+ "Model Precision": "float16",
751
+ "Model Parameters": "14.766B",
752
+ "Chat Template Use": "Yes"
753
+ },
754
+ {
755
+ "Model Rank": "28",
756
+ "Icon": "ð¬",
757
+ "Model Name": "Tsunami-th/Tsunami-1.0-14B-Instruct",
758
+ "Average Score": "41.84 %",
759
+ "IFEval Score": "78.29 %",
760
+ "BBH Score": "49.15 %",
761
+ "MATH Score": "45.85 %",
762
+ "GPQA Score": "14.21 %",
763
+ "MUSR Score": "16.34 %",
764
+ "MMLU-PRO Score": "47.21 %",
765
+ "Model Architecture": "Qwen2ForCausalLM",
766
+ "Model Precision": "bfloat16",
767
+ "Model Parameters": "14.77B",
768
+ "Chat Template Use": "Yes"
769
+ },
770
+ {
771
+ "Model Rank": "29",
772
+ "Icon": "├░┬╢",
773
+ "Model Name": "sthenno/tempesthenno-kto-0205-ckpt80",
774
+ "Average Score": "41.79 %",
775
+ "IFEval Score": "80.54 %",
776
+ "BBH Score": "50.64 %",
777
+ "MATH Score": "45.92 %",
778
+ "GPQA Score": "13.09 %",
779
+ "MUSR Score": "12.93 %",
780
+ "MMLU-PRO Score": "47.62 %",
781
+ "Model Architecture": "Qwen2ForCausalLM",
782
+ "Model Precision": "bfloat16",
783
+ "Model Parameters": "14.766B",
784
+ "Chat Template Use": "No"
785
+ },
786
+ {
787
+ "Model Rank": "30",
788
+ "Icon": "ð¤",
789
+ "Model Name": "sometimesanotion/Lamarck-14B-v0.7-rc4",
790
+ "Average Score": "41.79 %",
791
+ "IFEval Score": "72.11 %",
792
+ "BBH Score": "49.85 %",
793
+ "MATH Score": "40.26 %",
794
+ "GPQA Score": "18.57 %",
795
+ "MUSR Score": "21.07 %",
796
+ "MMLU-PRO Score": "48.89 %",
797
+ "Model Architecture": "Qwen2ForCausalLM",
798
+ "Model Precision": "bfloat16",
799
+ "Model Parameters": "14.766B",
800
+ "Chat Template Use": "No"
801
+ },
802
+ {
803
+ "Model Rank": "31",
804
+ "Icon": "├░┬╢",
805
+ "Model Name": "prithivMLmods/Porpoise-Opus-14B-Exp",
806
+ "Average Score": "41.77 %",
807
+ "IFEval Score": "70.98 %",
808
+ "BBH Score": "49.95 %",
809
+ "MATH Score": "40.41 %",
810
+ "GPQA Score": "19.13 %",
811
+ "MUSR Score": "21.30 %",
812
+ "MMLU-PRO Score": "48.85 %",
813
+ "Model Architecture": "Qwen2ForCausalLM",
814
+ "Model Precision": "bfloat16",
815
+ "Model Parameters": "14.766B",
816
+ "Chat Template Use": "No"
817
+ },
818
+ {
819
+ "Model Rank": "32",
820
+ "Icon": "ð¤",
821
+ "Model Name": "CombinHorizon/Josiefied-abliteratedV4-Qwen2.5-14B-Inst-BaseMerge-TIES",
822
+ "Average Score": "41.77 %",
823
+ "IFEval Score": "82.40 %",
824
+ "BBH Score": "48.20 %",
825
+ "MATH Score": "53.17 %",
826
+ "GPQA Score": "9.96 %",
827
+ "MUSR Score": "12.65 %",
828
+ "MMLU-PRO Score": "44.21 %",
829
+ "Model Architecture": "Qwen2ForCausalLM",
830
+ "Model Precision": "bfloat16",
831
+ "Model Parameters": "14.77B",
832
+ "Chat Template Use": "Yes"
833
+ },
834
  {
835
  "rank": 1,
836
  "name": "suayptalha/Lamarckvergence-14B",