Spaces:
Running
Running
LEADERBOARD_INTRODUCTION = """ | |
# DD-Ranking Leaderboard | |
</h3> --> | |
<p align="center"> | |
| <a href="https://nus-hpc-ai-lab.github.io/DD-Ranking/"><b>Documentation</b></a> | <a href="https://github.com/NUS-HPC-AI-Lab/DD-Ranking"><b>Github</b></a> | <a href=""><b>Paper </b> (Coming Soon)</a> | <a href=""><b>Twitter/X</b> (Coming Soon)</a> | <a href=""><b>Developer Slack</b> (Coming Soon)</a> | | |
</p> | |
🏆 Welcome to the leaderboard of the **DD-Ranking**! | |
> DD-Ranking (DD, i.e., Dataset Distillation) is an integrated and easy-to-use benchmark for dataset distillation. It aims to provide a fair evaluation scheme for DD methods that can decouple the impacts from knowledge distillation and data augmentation to reflect the real informativeness of the distilled data. | |
- **Fair Evaluation**: DD-Ranking provides a fair evaluation scheme for DD methods that can decouple the impacts from knowledge distillation and data augmentation to reflect the real informativeness of the distilled data. | |
- **Easy-to-use**: DD-Ranking provides a unified interface for dataset distillation evaluation. | |
- **Extensible**: DD-Ranking supports various datasets and models. | |
- **Customizable**: DD-Ranking supports various data augmentations and soft label strategies. | |
**Join Leaderboard**: Please see the [instructions](https://github.com/NUS-HPC-AI-Lab/DD-Ranking/blob/main/CONTRIBUTING.md) to participate. | |
""" | |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results" | |
CITATION_BUTTON_TEXT = r""" | |
COMING SOON | |
""" | |
IPC_INFO = """ | |
Images Per Class | |
""" | |
LABEL_TYPE_INFO = """ | |
Hard labels are categorical, having the same format of the real dataset. Soft labels are generated by a teacher model pretrained on the target dataset | |
""" | |
WEIGHT_ADJUSTMENT_INTRODUCTION = """ | |
The score for ranking in the following table is computed by $score = \sum w_i score_i$, where $w_i$ is the weight for the $i$-th metric. | |
**You can specify the weights for each metric below.** | |
""" | |
DATASET_LIST = ["CIFAR-10", "CIFAR-100", "Tiny-ImageNet"] | |
IPC_LIST = ["IPC-1", "IPC-10", "IPC-50"] | |
LABEL_TYPE_LIST = ["Hard Label", "Soft Label"] | |
METRICS = ["HLR", "IOR"] | |
COLUMN_NAMES = ["Ranking", "Method", "Verified", "Date", "Label Type", "HLR", "IOR", "Score"] | |
DATA_TITLE_TYPE = ['number', 'markdown', 'markdown', 'markdown', 'markdown', 'number', 'number', 'number'] | |
DATASET_MAPPING = { | |
"CIFAR-10": 0, | |
"CIFAR-100": 1, | |
"Tiny-ImageNet": 2, | |
} | |
IPC_MAPPING = { | |
"IPC-1": 0, | |
"IPC-10": 1, | |
"IPC-50": 2, | |
} | |
LABEL_MAPPING = { | |
"Hard Label": 0, | |
"Soft Label": 1, | |
} | |