--- dataset_info: features: - name: correct_answer dtype: string - name: incorrect_answer dtype: string - name: object dtype: string - name: original_image dtype: image - name: counterfact_image dtype: image splits: - name: color num_bytes: 260676475.2330435 num_examples: 573 - name: size num_bytes: 121141676.58352403 num_examples: 872 download_size: 379711088 dataset_size: 381818151.81656754 configs: - config_name: default data_files: - split: color path: data/color-* - split: size path: data/size-* task_categories: - image-text-to-text --- # Visual CounterFact: Controlling Knowledge Priors in Vision-Language Models through Visual Counterfactuals This dataset is part of the work **"Pixels Versus Priors: Controlling Knowledge Priors in Vision-Language Models through Visual Counterfacts"**. 📖 **[Read the Paper](https://arxiv.org/pdf/2505.17127)** 💾 **[GitHub Repository](https://github.com/rsinghlab/pixels_vs_priors)** ## Overview Visual CounterFact is a novel dataset designed to investigate how Multimodal Large Language Models (MLLMs) balance memorized world knowledge priors (e.g., "strawberries are red") with the visual evidence present in input images (e.g., a blue strawberry). The dataset features visually realistic counterfactuals that create direct conflicts between what the model has learned and what it sees. This allows for studying and controlling whether model predictions rely on memorized priors or the actual image content. ## Dataset Splits The dataset contains two distinct splits, each corresponding to a specific visual attribute reasoning task: ### **Color (`color`)** - **Description:** Contains images of objects where the color attribute is either consistent with common world knowledge or is a counterfactual color designed to contradict it (e.g., a blue strawberry). - **Purpose:** Evaluate how models reconcile conflicting color information between prior knowledge and visual input. - **Example Queries:** - *"What color is this strawberry?"* - *"What color are most strawberries?"* ### **Size (`size`)** - **Description:** Consists of object images with size relations manipulated to contradict typical real-world size expectations (e.g., a fly larger than a strawberry). - **Purpose:** Test model understanding of size priors versus visual evidence. - **Example Queries:** - *"Which object is bigger in this image, the fly or the strawberry?"* - *"Are strawberries bigger than flies?"* ## Citation If you use this dataset, please cite: > Pixels Versus Priors: Controlling Knowledge Priors in Vision-Language Models through Visual Counterfacts