Spaces:
Running
Running
Update constants.py
Browse files- constants.py +19 -3
constants.py
CHANGED
@@ -59,10 +59,26 @@ Hard labels are categorical, having the same format of the real dataset. Soft la
|
|
59 |
"""
|
60 |
|
61 |
WEIGHT_ADJUSTMENT_INTRODUCTION = """
|
62 |
-
The score for ranking in the following table is computed by
|
63 |
**You can specify the weight $w$ below.**
|
64 |
"""
|
65 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
66 |
DATASET_LIST = ["CIFAR-10", "CIFAR-100", "Tiny-ImageNet"]
|
67 |
IPC_LIST = ["IPC-1", "IPC-10", "IPC-50"]
|
68 |
DATASET_IPC_LIST = {
|
@@ -74,7 +90,7 @@ LABEL_TYPE_LIST = ["Hard Label", "Soft Label"]
|
|
74 |
|
75 |
METRICS = ["HLR", "IOR"]
|
76 |
METRICS_SIGN = [1.0, -1.0]
|
77 |
-
COLUMN_NAMES = ["Ranking", "Method", "Verified", "Date", "Label Type", "HLR", "IOR", "
|
78 |
DATA_TITLE_TYPE = ['number', 'markdown', 'markdown', 'markdown', 'markdown', 'number', 'number', 'number']
|
79 |
|
80 |
DATASET_MAPPING = {
|
@@ -92,4 +108,4 @@ IPC_MAPPING = {
|
|
92 |
LABEL_MAPPING = {
|
93 |
"Hard Label": 0,
|
94 |
"Soft Label": 1,
|
95 |
-
}
|
|
|
59 |
"""
|
60 |
|
61 |
WEIGHT_ADJUSTMENT_INTRODUCTION = """
|
62 |
+
The score for ranking (DD-Ranking Score, DDRS) in the following table is computed by DDRS = \\frac{e^{w IOR - (1 - w) HLR} - e^{-1}}{e - e^{-1}}$, where $w$ is the weight for the HLR metric.
|
63 |
**You can specify the weight $w$ below.**
|
64 |
"""
|
65 |
|
66 |
+
METRIC_DEFINITION_INTRODUCTION = """
|
67 |
+
$\\text{Acc.}$: The accuracy of models trained on different samples.
|
68 |
+
|
69 |
+
$\\text{full-hard}$: Full dataset with hard labels.
|
70 |
+
|
71 |
+
$\\text{syn-hard}$: Synthetic dataset with hard labels.
|
72 |
+
|
73 |
+
$\\text{syn-any}$: Synthetic dataset with personalized evaluation methods (hard or soft labels).
|
74 |
+
|
75 |
+
$\\text{rdm-any}$: Randomly selected dataset (under the same compression ratio) with the same personalized evaluation methods.
|
76 |
+
|
77 |
+
$\\text{HLR} = \\text{Acc.} \\text{full-hard} - \\text{Acc.} \\text{syn-hard}$: The degree to which the original dataset is recovered under hard labels (hard label recovery).
|
78 |
+
|
79 |
+
$\\text{IOR} = \\text{Acc.} \\text{syn-any} - \\text{Acc.} \\text{rdm-any}$: The improvement over random selection when using personalized evaluation methods (improvement over random).
|
80 |
+
"""
|
81 |
+
|
82 |
DATASET_LIST = ["CIFAR-10", "CIFAR-100", "Tiny-ImageNet"]
|
83 |
IPC_LIST = ["IPC-1", "IPC-10", "IPC-50"]
|
84 |
DATASET_IPC_LIST = {
|
|
|
90 |
|
91 |
METRICS = ["HLR", "IOR"]
|
92 |
METRICS_SIGN = [1.0, -1.0]
|
93 |
+
COLUMN_NAMES = ["Ranking", "Method", "Verified", "Date", "Label Type", "HLR%", "IOR%", "DDRS"]
|
94 |
DATA_TITLE_TYPE = ['number', 'markdown', 'markdown', 'markdown', 'markdown', 'number', 'number', 'number']
|
95 |
|
96 |
DATASET_MAPPING = {
|
|
|
108 |
LABEL_MAPPING = {
|
109 |
"Hard Label": 0,
|
110 |
"Soft Label": 1,
|
111 |
+
}
|