harpreetsahota commited on
Commit
8ffb3b7
·
verified ·
1 Parent(s): ef1f67d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -5
README.md CHANGED
@@ -82,16 +82,14 @@ session = fo.launch_app(dataset)
82
 
83
  ### Dataset Description
84
 
85
- ImageNet-D is a new benchmark created using diffusion models to generate realistic synthetic images with diverse backgrounds, textures, and materials[1]. The dataset contains 4,835 hard images that cause significant accuracy drops of up to 60% for a range of vision models, including ResNet, ViT, CLIP, LLaVa, and MiniGPT-4[1].
86
 
87
- To create ImageNet-D, a large pool of synthetic images is generated by combining object categories with various nuisance attributes using Stable Diffusion[1]. The most challenging images that cause shared failures across multiple surrogate models are selected for the final dataset[1]. Human labelling via Amazon Mechanical Turk is used for quality control to ensure the images are valid and high-quality[1].
88
 
89
- Experiments show that ImageNet-D reveals significant robustness gaps in current vision models[1]. The synthetic images transfer well to unseen models, uncovering common failure modes[1]. ImageNet-D provides a more diverse and challenging test set than prior synthetic benchmarks like ImageNet-C, ImageNet-9, and Stylized ImageNet[1].
90
 
91
  The recipe notebook for creating this dataset can be found [here](https://colab.research.google.com/drive/1iiiXN8B36YhjtOH2PDbHevHTXH736It_?usp=sharing)
92
 
93
- Citations:
94
- [1] https://arxiv.org/html/2403.18775v1
95
 
96
  - **Funded by :** KAIST, University of Michigan, Ann Arbor, McGill University, MILA
97
  - **License:** MIT License
 
82
 
83
  ### Dataset Description
84
 
85
+ ImageNet-D is a new benchmark created using diffusion models to generate realistic synthetic images with diverse backgrounds, textures, and materials. The dataset contains 4,835 hard images that cause significant accuracy drops of up to 60% for a range of vision models, including ResNet, ViT, CLIP, LLaVa, and MiniGPT-4.
86
 
87
+ To create ImageNet-D, a large pool of synthetic images is generated by combining object categories with various nuisance attributes using Stable Diffusion. The most challenging images that cause shared failures across multiple surrogate models are selected for the final dataset. Human labelling via Amazon Mechanical Turk is used for quality control to ensure the images are valid and high-quality.
88
 
89
+ Experiments show that ImageNet-D reveals significant robustness gaps in current vision models. The synthetic images transfer well to unseen models, uncovering common failure modes. ImageNet-D provides a more diverse and challenging test set than prior synthetic benchmarks like ImageNet-C, ImageNet-9, and Stylized ImageNet.
90
 
91
  The recipe notebook for creating this dataset can be found [here](https://colab.research.google.com/drive/1iiiXN8B36YhjtOH2PDbHevHTXH736It_?usp=sharing)
92
 
 
 
93
 
94
  - **Funded by :** KAIST, University of Michigan, Ann Arbor, McGill University, MILA
95
  - **License:** MIT License