Upload Qwen/Qwen2.5-14B/results_2025-01-19T10-31-14.129542.json with huggingface_hub
Browse files
Qwen/Qwen2.5-14B/results_2025-01-19T10-31-14.129542.json
CHANGED
@@ -451,36 +451,36 @@
|
|
451 |
"acc_norm_stderr": 0.02537956363744994
|
452 |
},
|
453 |
"community|aratrust:Ethics|0": {
|
454 |
-
"
|
455 |
-
"
|
456 |
},
|
457 |
"community|aratrust:Illegal|0": {
|
458 |
-
"
|
459 |
-
"
|
460 |
},
|
461 |
"community|aratrust:MentalHealth|0": {
|
462 |
-
"
|
463 |
-
"
|
464 |
},
|
465 |
"community|aratrust:Offensive|0": {
|
466 |
-
"
|
467 |
-
"
|
468 |
},
|
469 |
"community|aratrust:PhysicalHealth|0": {
|
470 |
-
"
|
471 |
-
"
|
472 |
},
|
473 |
"community|aratrust:Privacy|0": {
|
474 |
-
"
|
475 |
-
"
|
476 |
},
|
477 |
"community|aratrust:Trustfulness|0": {
|
478 |
-
"
|
479 |
-
"
|
480 |
},
|
481 |
"community|aratrust:Unfairness|0": {
|
482 |
-
"
|
483 |
-
"
|
484 |
},
|
485 |
"community|alghafa:_average|0": {
|
486 |
"acc_norm": 0.4993301141153749,
|
@@ -499,14 +499,18 @@
|
|
499 |
"acc_norm_stderr": 0.02257288781426151
|
500 |
},
|
501 |
"community|aratrust:_average|0": {
|
502 |
-
"
|
503 |
-
"
|
504 |
},
|
505 |
"all": {
|
506 |
"acc_norm": 0.492983346222567,
|
507 |
"acc_norm_stderr": 0.03327366955636232,
|
508 |
"f1": 0.011369886122049063,
|
509 |
"f1_stderr": 0.006846190218550371
|
|
|
|
|
|
|
|
|
510 |
}
|
511 |
},
|
512 |
"versions": {
|
@@ -626,7 +630,8 @@
|
|
626 |
"community|aratrust:Trustfulness|0": 0,
|
627 |
"community|aratrust:Unfairness|0": 0,
|
628 |
"community|madinah_qa:Arabic Language (General)|0": 0,
|
629 |
-
"community|madinah_qa:Arabic Language (Grammar)|0": 0
|
|
|
630 |
},
|
631 |
"config_tasks": {
|
632 |
"community|alghafa:mcq_exams_test_ar": {
|
@@ -4792,9 +4797,9 @@
|
|
4792 |
"hf_subset": "Ethics",
|
4793 |
"metric": [
|
4794 |
{
|
4795 |
-
"metric_name": "
|
4796 |
"higher_is_better": true,
|
4797 |
-
"category": "
|
4798 |
"use_case": "1",
|
4799 |
"sample_level_fn": "compute",
|
4800 |
"corpus_level_fn": "mean"
|
@@ -4830,9 +4835,9 @@
|
|
4830 |
"hf_subset": "Illegal",
|
4831 |
"metric": [
|
4832 |
{
|
4833 |
-
"metric_name": "
|
4834 |
"higher_is_better": true,
|
4835 |
-
"category": "
|
4836 |
"use_case": "1",
|
4837 |
"sample_level_fn": "compute",
|
4838 |
"corpus_level_fn": "mean"
|
@@ -4868,9 +4873,9 @@
|
|
4868 |
"hf_subset": "MentalHealth",
|
4869 |
"metric": [
|
4870 |
{
|
4871 |
-
"metric_name": "
|
4872 |
"higher_is_better": true,
|
4873 |
-
"category": "
|
4874 |
"use_case": "1",
|
4875 |
"sample_level_fn": "compute",
|
4876 |
"corpus_level_fn": "mean"
|
@@ -4906,9 +4911,9 @@
|
|
4906 |
"hf_subset": "Offensive",
|
4907 |
"metric": [
|
4908 |
{
|
4909 |
-
"metric_name": "
|
4910 |
"higher_is_better": true,
|
4911 |
-
"category": "
|
4912 |
"use_case": "1",
|
4913 |
"sample_level_fn": "compute",
|
4914 |
"corpus_level_fn": "mean"
|
@@ -4944,9 +4949,9 @@
|
|
4944 |
"hf_subset": "PhysicalHealth",
|
4945 |
"metric": [
|
4946 |
{
|
4947 |
-
"metric_name": "
|
4948 |
"higher_is_better": true,
|
4949 |
-
"category": "
|
4950 |
"use_case": "1",
|
4951 |
"sample_level_fn": "compute",
|
4952 |
"corpus_level_fn": "mean"
|
@@ -4982,9 +4987,9 @@
|
|
4982 |
"hf_subset": "Privacy",
|
4983 |
"metric": [
|
4984 |
{
|
4985 |
-
"metric_name": "
|
4986 |
"higher_is_better": true,
|
4987 |
-
"category": "
|
4988 |
"use_case": "1",
|
4989 |
"sample_level_fn": "compute",
|
4990 |
"corpus_level_fn": "mean"
|
@@ -5020,9 +5025,9 @@
|
|
5020 |
"hf_subset": "Trustfulness",
|
5021 |
"metric": [
|
5022 |
{
|
5023 |
-
"metric_name": "
|
5024 |
"higher_is_better": true,
|
5025 |
-
"category": "
|
5026 |
"use_case": "1",
|
5027 |
"sample_level_fn": "compute",
|
5028 |
"corpus_level_fn": "mean"
|
@@ -5058,9 +5063,9 @@
|
|
5058 |
"hf_subset": "Unfairness",
|
5059 |
"metric": [
|
5060 |
{
|
5061 |
-
"metric_name": "
|
5062 |
"higher_is_better": true,
|
5063 |
-
"category": "
|
5064 |
"use_case": "1",
|
5065 |
"sample_level_fn": "compute",
|
5066 |
"corpus_level_fn": "mean"
|
@@ -5168,6 +5173,44 @@
|
|
5168 |
"effective_num_docs": 365,
|
5169 |
"must_remove_duplicate_docs": false,
|
5170 |
"version": 0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5171 |
}
|
5172 |
},
|
5173 |
"summary_tasks": {
|
@@ -6699,113 +6742,127 @@
|
|
6699 |
},
|
6700 |
"community|aratrust:Ethics|0": {
|
6701 |
"hashes": {
|
6702 |
-
"hash_examples": "
|
6703 |
-
"hash_full_prompts": "
|
6704 |
-
"hash_input_tokens": "
|
6705 |
-
"hash_cont_tokens": "
|
6706 |
},
|
6707 |
-
"truncated":
|
6708 |
-
"non_truncated":
|
6709 |
-
"padded":
|
6710 |
"non_padded": 0,
|
6711 |
"effective_few_shots": 0.0,
|
6712 |
"num_truncated_few_shots": 0
|
6713 |
},
|
6714 |
"community|aratrust:Illegal|0": {
|
6715 |
"hashes": {
|
6716 |
-
"hash_examples": "
|
6717 |
-
"hash_full_prompts": "
|
6718 |
-
"hash_input_tokens": "
|
6719 |
-
"hash_cont_tokens": "
|
6720 |
},
|
6721 |
-
"truncated":
|
6722 |
-
"non_truncated":
|
6723 |
-
"padded":
|
6724 |
"non_padded": 0,
|
6725 |
"effective_few_shots": 0.0,
|
6726 |
"num_truncated_few_shots": 0
|
6727 |
},
|
6728 |
"community|aratrust:MentalHealth|0": {
|
6729 |
"hashes": {
|
6730 |
-
"hash_examples": "
|
6731 |
-
"hash_full_prompts": "
|
6732 |
-
"hash_input_tokens": "
|
6733 |
-
"hash_cont_tokens": "
|
6734 |
},
|
6735 |
-
"truncated":
|
6736 |
-
"non_truncated":
|
6737 |
-
"padded":
|
6738 |
"non_padded": 0,
|
6739 |
"effective_few_shots": 0.0,
|
6740 |
"num_truncated_few_shots": 0
|
6741 |
},
|
6742 |
"community|aratrust:Offensive|0": {
|
6743 |
"hashes": {
|
6744 |
-
"hash_examples": "
|
6745 |
-
"hash_full_prompts": "
|
6746 |
-
"hash_input_tokens": "
|
6747 |
-
"hash_cont_tokens": "
|
6748 |
},
|
6749 |
-
"truncated":
|
6750 |
-
"non_truncated":
|
6751 |
-
"padded":
|
6752 |
"non_padded": 0,
|
6753 |
"effective_few_shots": 0.0,
|
6754 |
"num_truncated_few_shots": 0
|
6755 |
},
|
6756 |
"community|aratrust:PhysicalHealth|0": {
|
6757 |
"hashes": {
|
6758 |
-
"hash_examples": "
|
6759 |
-
"hash_full_prompts": "
|
6760 |
-
"hash_input_tokens": "
|
6761 |
-
"hash_cont_tokens": "
|
6762 |
},
|
6763 |
-
"truncated":
|
6764 |
-
"non_truncated":
|
6765 |
-
"padded":
|
6766 |
-
"non_padded":
|
6767 |
"effective_few_shots": 0.0,
|
6768 |
"num_truncated_few_shots": 0
|
6769 |
},
|
6770 |
"community|aratrust:Privacy|0": {
|
6771 |
"hashes": {
|
6772 |
-
"hash_examples": "
|
6773 |
-
"hash_full_prompts": "
|
6774 |
-
"hash_input_tokens": "
|
6775 |
-
"hash_cont_tokens": "
|
6776 |
},
|
6777 |
-
"truncated":
|
6778 |
-
"non_truncated":
|
6779 |
-
"padded":
|
6780 |
-
"non_padded":
|
6781 |
"effective_few_shots": 0.0,
|
6782 |
"num_truncated_few_shots": 0
|
6783 |
},
|
6784 |
"community|aratrust:Trustfulness|0": {
|
6785 |
"hashes": {
|
6786 |
-
"hash_examples": "
|
6787 |
-
"hash_full_prompts": "
|
6788 |
-
"hash_input_tokens": "
|
6789 |
-
"hash_cont_tokens": "
|
6790 |
},
|
6791 |
"truncated": 0,
|
6792 |
"non_truncated": 78,
|
6793 |
-
"padded":
|
6794 |
-
"non_padded":
|
6795 |
"effective_few_shots": 0.0,
|
6796 |
"num_truncated_few_shots": 0
|
6797 |
},
|
6798 |
"community|aratrust:Unfairness|0": {
|
6799 |
"hashes": {
|
6800 |
-
"hash_examples": "
|
6801 |
-
"hash_full_prompts": "
|
6802 |
-
"hash_input_tokens": "
|
6803 |
-
"hash_cont_tokens": "
|
6804 |
},
|
6805 |
-
"truncated":
|
6806 |
-
"non_truncated":
|
6807 |
-
"padded":
|
6808 |
-
"non_padded":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6809 |
"effective_few_shots": 0.0,
|
6810 |
"num_truncated_few_shots": 0
|
6811 |
}
|
|
|
451 |
"acc_norm_stderr": 0.02537956363744994
|
452 |
},
|
453 |
"community|aratrust:Ethics|0": {
|
454 |
+
"f1": 0.0,
|
455 |
+
"f1_stderr": 0.0
|
456 |
},
|
457 |
"community|aratrust:Illegal|0": {
|
458 |
+
"f1": 0.0,
|
459 |
+
"f1_stderr": 0.0
|
460 |
},
|
461 |
"community|aratrust:MentalHealth|0": {
|
462 |
+
"f1": 0.013157894736842105,
|
463 |
+
"f1_stderr": 0.013157894736842105
|
464 |
},
|
465 |
"community|aratrust:Offensive|0": {
|
466 |
+
"f1": 0.0,
|
467 |
+
"f1_stderr": 0.0
|
468 |
},
|
469 |
"community|aratrust:PhysicalHealth|0": {
|
470 |
+
"f1": 0.0136986301369863,
|
471 |
+
"f1_stderr": 0.013698630136986313
|
472 |
},
|
473 |
"community|aratrust:Privacy|0": {
|
474 |
+
"f1": 0.0,
|
475 |
+
"f1_stderr": 0.0
|
476 |
},
|
477 |
"community|aratrust:Trustfulness|0": {
|
478 |
+
"f1": 0.0641025641025641,
|
479 |
+
"f1_stderr": 0.027912996874574544
|
480 |
},
|
481 |
"community|aratrust:Unfairness|0": {
|
482 |
+
"f1": 0.0,
|
483 |
+
"f1_stderr": 0.0
|
484 |
},
|
485 |
"community|alghafa:_average|0": {
|
486 |
"acc_norm": 0.4993301141153749,
|
|
|
499 |
"acc_norm_stderr": 0.02257288781426151
|
500 |
},
|
501 |
"community|aratrust:_average|0": {
|
502 |
+
"f1": 0.011369886122049063,
|
503 |
+
"f1_stderr": 0.006846190218550371
|
504 |
},
|
505 |
"all": {
|
506 |
"acc_norm": 0.492983346222567,
|
507 |
"acc_norm_stderr": 0.03327366955636232,
|
508 |
"f1": 0.011369886122049063,
|
509 |
"f1_stderr": 0.006846190218550371
|
510 |
+
},
|
511 |
+
"community|alrage_qa|0": {
|
512 |
+
"llm_as_judge": 0.7402659069325689,
|
513 |
+
"llm_as_judge_stderr": 0.00013976642991211927
|
514 |
}
|
515 |
},
|
516 |
"versions": {
|
|
|
630 |
"community|aratrust:Trustfulness|0": 0,
|
631 |
"community|aratrust:Unfairness|0": 0,
|
632 |
"community|madinah_qa:Arabic Language (General)|0": 0,
|
633 |
+
"community|madinah_qa:Arabic Language (Grammar)|0": 0,
|
634 |
+
"community|alrage_qa|0": 0
|
635 |
},
|
636 |
"config_tasks": {
|
637 |
"community|alghafa:mcq_exams_test_ar": {
|
|
|
4797 |
"hf_subset": "Ethics",
|
4798 |
"metric": [
|
4799 |
{
|
4800 |
+
"metric_name": "f1",
|
4801 |
"higher_is_better": true,
|
4802 |
+
"category": "3",
|
4803 |
"use_case": "1",
|
4804 |
"sample_level_fn": "compute",
|
4805 |
"corpus_level_fn": "mean"
|
|
|
4835 |
"hf_subset": "Illegal",
|
4836 |
"metric": [
|
4837 |
{
|
4838 |
+
"metric_name": "f1",
|
4839 |
"higher_is_better": true,
|
4840 |
+
"category": "3",
|
4841 |
"use_case": "1",
|
4842 |
"sample_level_fn": "compute",
|
4843 |
"corpus_level_fn": "mean"
|
|
|
4873 |
"hf_subset": "MentalHealth",
|
4874 |
"metric": [
|
4875 |
{
|
4876 |
+
"metric_name": "f1",
|
4877 |
"higher_is_better": true,
|
4878 |
+
"category": "3",
|
4879 |
"use_case": "1",
|
4880 |
"sample_level_fn": "compute",
|
4881 |
"corpus_level_fn": "mean"
|
|
|
4911 |
"hf_subset": "Offensive",
|
4912 |
"metric": [
|
4913 |
{
|
4914 |
+
"metric_name": "f1",
|
4915 |
"higher_is_better": true,
|
4916 |
+
"category": "3",
|
4917 |
"use_case": "1",
|
4918 |
"sample_level_fn": "compute",
|
4919 |
"corpus_level_fn": "mean"
|
|
|
4949 |
"hf_subset": "PhysicalHealth",
|
4950 |
"metric": [
|
4951 |
{
|
4952 |
+
"metric_name": "f1",
|
4953 |
"higher_is_better": true,
|
4954 |
+
"category": "3",
|
4955 |
"use_case": "1",
|
4956 |
"sample_level_fn": "compute",
|
4957 |
"corpus_level_fn": "mean"
|
|
|
4987 |
"hf_subset": "Privacy",
|
4988 |
"metric": [
|
4989 |
{
|
4990 |
+
"metric_name": "f1",
|
4991 |
"higher_is_better": true,
|
4992 |
+
"category": "3",
|
4993 |
"use_case": "1",
|
4994 |
"sample_level_fn": "compute",
|
4995 |
"corpus_level_fn": "mean"
|
|
|
5025 |
"hf_subset": "Trustfulness",
|
5026 |
"metric": [
|
5027 |
{
|
5028 |
+
"metric_name": "f1",
|
5029 |
"higher_is_better": true,
|
5030 |
+
"category": "3",
|
5031 |
"use_case": "1",
|
5032 |
"sample_level_fn": "compute",
|
5033 |
"corpus_level_fn": "mean"
|
|
|
5063 |
"hf_subset": "Unfairness",
|
5064 |
"metric": [
|
5065 |
{
|
5066 |
+
"metric_name": "f1",
|
5067 |
"higher_is_better": true,
|
5068 |
+
"category": "3",
|
5069 |
"use_case": "1",
|
5070 |
"sample_level_fn": "compute",
|
5071 |
"corpus_level_fn": "mean"
|
|
|
5173 |
"effective_num_docs": 365,
|
5174 |
"must_remove_duplicate_docs": false,
|
5175 |
"version": 0
|
5176 |
+
},
|
5177 |
+
"community|alrage_qa": {
|
5178 |
+
"name": "alrage_qa",
|
5179 |
+
"prompt_function": "qa_prompt_arabic",
|
5180 |
+
"hf_repo": "OALL/ALRAGE",
|
5181 |
+
"hf_subset": null,
|
5182 |
+
"metric": [
|
5183 |
+
{
|
5184 |
+
"metric_name": "llm_as_judge",
|
5185 |
+
"higher_is_better": true,
|
5186 |
+
"category": "7",
|
5187 |
+
"use_case": "10",
|
5188 |
+
"sample_level_fn": "_sample_level_fn",
|
5189 |
+
"corpus_level_fn": "aggregate_scores"
|
5190 |
+
}
|
5191 |
+
],
|
5192 |
+
"hf_revision": null,
|
5193 |
+
"hf_filter": null,
|
5194 |
+
"hf_avail_splits": [
|
5195 |
+
"train"
|
5196 |
+
],
|
5197 |
+
"trust_dataset": true,
|
5198 |
+
"evaluation_splits": [
|
5199 |
+
"train"
|
5200 |
+
],
|
5201 |
+
"few_shots_split": null,
|
5202 |
+
"few_shots_select": null,
|
5203 |
+
"generation_size": 200,
|
5204 |
+
"generation_grammar": null,
|
5205 |
+
"stop_sequence": [],
|
5206 |
+
"num_samples": null,
|
5207 |
+
"suite": [
|
5208 |
+
"community"
|
5209 |
+
],
|
5210 |
+
"original_num_docs": 2106,
|
5211 |
+
"effective_num_docs": 2106,
|
5212 |
+
"must_remove_duplicate_docs": false,
|
5213 |
+
"version": 0
|
5214 |
}
|
5215 |
},
|
5216 |
"summary_tasks": {
|
|
|
6742 |
},
|
6743 |
"community|aratrust:Ethics|0": {
|
6744 |
"hashes": {
|
6745 |
+
"hash_examples": "b77354655caca219",
|
6746 |
+
"hash_full_prompts": "6454a804fd0d7e23",
|
6747 |
+
"hash_input_tokens": "c0bf39e0a72f740f",
|
6748 |
+
"hash_cont_tokens": "fb652402ea847f88"
|
6749 |
},
|
6750 |
+
"truncated": 16,
|
6751 |
+
"non_truncated": 44,
|
6752 |
+
"padded": 60,
|
6753 |
"non_padded": 0,
|
6754 |
"effective_few_shots": 0.0,
|
6755 |
"num_truncated_few_shots": 0
|
6756 |
},
|
6757 |
"community|aratrust:Illegal|0": {
|
6758 |
"hashes": {
|
6759 |
+
"hash_examples": "daa90cfb03dd9ed8",
|
6760 |
+
"hash_full_prompts": "07ba46f20165ca3a",
|
6761 |
+
"hash_input_tokens": "029636665eee2841",
|
6762 |
+
"hash_cont_tokens": "c59c5f2d30dbefe6"
|
6763 |
},
|
6764 |
+
"truncated": 22,
|
6765 |
+
"non_truncated": 31,
|
6766 |
+
"padded": 53,
|
6767 |
"non_padded": 0,
|
6768 |
"effective_few_shots": 0.0,
|
6769 |
"num_truncated_few_shots": 0
|
6770 |
},
|
6771 |
"community|aratrust:MentalHealth|0": {
|
6772 |
"hashes": {
|
6773 |
+
"hash_examples": "ca046355c96d95d9",
|
6774 |
+
"hash_full_prompts": "68e35c11be14dd61",
|
6775 |
+
"hash_input_tokens": "85e5e6d35b135d4f",
|
6776 |
+
"hash_cont_tokens": "1769d2f73f2da43f"
|
6777 |
},
|
6778 |
+
"truncated": 48,
|
6779 |
+
"non_truncated": 28,
|
6780 |
+
"padded": 76,
|
6781 |
"non_padded": 0,
|
6782 |
"effective_few_shots": 0.0,
|
6783 |
"num_truncated_few_shots": 0
|
6784 |
},
|
6785 |
"community|aratrust:Offensive|0": {
|
6786 |
"hashes": {
|
6787 |
+
"hash_examples": "6ff77d23c0f3113d",
|
6788 |
+
"hash_full_prompts": "8965fc596a628547",
|
6789 |
+
"hash_input_tokens": "52a605d2d22e5afd",
|
6790 |
+
"hash_cont_tokens": "af9ca888e4a63bdf"
|
6791 |
},
|
6792 |
+
"truncated": 21,
|
6793 |
+
"non_truncated": 48,
|
6794 |
+
"padded": 69,
|
6795 |
"non_padded": 0,
|
6796 |
"effective_few_shots": 0.0,
|
6797 |
"num_truncated_few_shots": 0
|
6798 |
},
|
6799 |
"community|aratrust:PhysicalHealth|0": {
|
6800 |
"hashes": {
|
6801 |
+
"hash_examples": "085db2421f8abf29",
|
6802 |
+
"hash_full_prompts": "e90845039dc9435f",
|
6803 |
+
"hash_input_tokens": "afab808390420e65",
|
6804 |
+
"hash_cont_tokens": "ae2c45e101dfdaa8"
|
6805 |
},
|
6806 |
+
"truncated": 51,
|
6807 |
+
"non_truncated": 22,
|
6808 |
+
"padded": 73,
|
6809 |
+
"non_padded": 0,
|
6810 |
"effective_few_shots": 0.0,
|
6811 |
"num_truncated_few_shots": 0
|
6812 |
},
|
6813 |
"community|aratrust:Privacy|0": {
|
6814 |
"hashes": {
|
6815 |
+
"hash_examples": "78f4d16753b18c49",
|
6816 |
+
"hash_full_prompts": "e28b2c6a166889d2",
|
6817 |
+
"hash_input_tokens": "5366ac73b2237429",
|
6818 |
+
"hash_cont_tokens": "aa9be218b39168d8"
|
6819 |
},
|
6820 |
+
"truncated": 36,
|
6821 |
+
"non_truncated": 21,
|
6822 |
+
"padded": 56,
|
6823 |
+
"non_padded": 1,
|
6824 |
"effective_few_shots": 0.0,
|
6825 |
"num_truncated_few_shots": 0
|
6826 |
},
|
6827 |
"community|aratrust:Trustfulness|0": {
|
6828 |
"hashes": {
|
6829 |
+
"hash_examples": "373f72b4e30243c4",
|
6830 |
+
"hash_full_prompts": "401025b282b044ac",
|
6831 |
+
"hash_input_tokens": "628a5c75ae9e166a",
|
6832 |
+
"hash_cont_tokens": "9977283a343f95e8"
|
6833 |
},
|
6834 |
"truncated": 0,
|
6835 |
"non_truncated": 78,
|
6836 |
+
"padded": 78,
|
6837 |
+
"non_padded": 0,
|
6838 |
"effective_few_shots": 0.0,
|
6839 |
"num_truncated_few_shots": 0
|
6840 |
},
|
6841 |
"community|aratrust:Unfairness|0": {
|
6842 |
"hashes": {
|
6843 |
+
"hash_examples": "51fa7940e42ffcc6",
|
6844 |
+
"hash_full_prompts": "665e1c47ce0fe4c0",
|
6845 |
+
"hash_input_tokens": "e4b7ee5e0dad6b5a",
|
6846 |
+
"hash_cont_tokens": "9231ac74de2c9c55"
|
6847 |
},
|
6848 |
+
"truncated": 25,
|
6849 |
+
"non_truncated": 30,
|
6850 |
+
"padded": 55,
|
6851 |
+
"non_padded": 0,
|
6852 |
+
"effective_few_shots": 0.0,
|
6853 |
+
"num_truncated_few_shots": 0
|
6854 |
+
},
|
6855 |
+
"community|alrage_qa|0": {
|
6856 |
+
"hashes": {
|
6857 |
+
"hash_examples": "3edbbe22cabd4160",
|
6858 |
+
"hash_full_prompts": "91c8d5ed9f9796ea",
|
6859 |
+
"hash_input_tokens": "338c70b76f6963cd",
|
6860 |
+
"hash_cont_tokens": "5e63b4eb64ad6499"
|
6861 |
+
},
|
6862 |
+
"truncated": 2106,
|
6863 |
+
"non_truncated": 0,
|
6864 |
+
"padded": 2106,
|
6865 |
+
"non_padded": 0,
|
6866 |
"effective_few_shots": 0.0,
|
6867 |
"num_truncated_few_shots": 0
|
6868 |
}
|