File size: 4,146 Bytes
1b961e9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cb28376
1b961e9
cb28376
1b961e9
 
 
 
 
cb28376
1b961e9
cb28376
 
 
 
 
1b961e9
cb28376
1b961e9
cb28376
 
1b961e9
 
 
 
 
 
cb28376
1b961e9
 
 
cb28376
1b961e9
 
 
cb28376
1b961e9
 
 
cb28376
1b961e9
 
 
cb28376
1b961e9
 
 
cb28376
1b961e9
 
 
 
 
 
cb28376
1b961e9
 
 
cb28376
1b961e9
 
 
cb28376
1b961e9
cb28376
1b961e9
cb28376
2a3e3e7
 
 
 
 
 
 
 
 
cb28376
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
---
base_model: Qwen/Qwen2-VL-7B-Instruct
datasets:
- aliencaocao/multimodal_meme_classification_singapore
language: en
library_name: transformers
license: mit
metrics:
- accuracy
- roc_auc
pipeline_tag: image-text-to-text
tags:
- memes
- offensive
- singapore
- vlm
model-index:
- name: qwen2-vl-7b-rslora-offensive-meme-singapore
  results:
  - task:
      type: image-classification
      name: Offensive Meme Classification
    dataset:
      name: Offensive Memes in Singapore Context
      type: aliencaocao/multimodal_meme_classification_singapore
      split: test
    metrics:
    - type: roc_auc
      value: 0.8192
      name: AUROC
    - type: accuracy
      value: 0.8043
      name: Accuracy
---

# Model Card for Qwen2-VL 7B RSLORA Offensive Meme Singapore

This model is a fine-tuned version of Qwen2-VL-7B-Instruct for offensive meme classification in the Singapore context.  It was trained on the [multimodal_meme_classification_singapore](https://huggingface.co/datasets/aliencaocao/multimodal_meme_classification_singapore) dataset.

## Model Details

### Model Description

This model classifies memes as offensive or not, taking into account Singaporean social context. It leverages the visual and textual understanding capabilities of Qwen2-VL-7B-Instruct.

- **Developed by:** Cao Yuxuan, Wu Jiayang, Alistair Cheong Liang Chuen, Bryan Shan Guanrong, Theodore Lee Chong Jen, and Sherman Chann Zhi Shen
- **Model type:** Vision-Language Model (VLM)
- **Language(s) (NLP):** en
- **License:** MIT
- **Finetuned from model:** Qwen/Qwen2-VL-7B-Instruct

### Model Sources

- **Repository:** [https://github.com/aliencaocao/vlm-for-memes-aisg](https://github.com/aliencaocao/vlm-for-memes-aisg)
- **Paper:** [Detecting Offensive Memes with Social Biases in Singapore Context Using Multimodal Large Language Models](https://arxiv.org/abs/2502.18101)


## Uses

### Direct Use

This model can be used directly to classify memes.  See the code example in the "How to Get Started" section.

### Downstream Use [optional]

This model can be further fine-tuned for other related tasks or incorporated into a larger content moderation system.

### Out-of-Scope Use

This model is specifically trained for the Singaporean context and may not generalize well to other cultures or languages.  It should not be used to make definitive judgments about individuals or groups.

## Bias, Risks, and Limitations

Like any machine learning model, this model may exhibit biases present in the training data.  It is important to be aware of these limitations and use the model responsibly.  Further research is needed to assess and mitigate potential biases.

### Recommendations

Users should be aware of the potential for bias and limitations in the model's performance.  It is recommended to use this model as a tool to assist human moderators rather than a replacement for human judgment.

## How to Get Started with the Model

See the model repository's README for usage examples: [https://github.com/aliencaocao/vlm-for-memes-aisg](https://github.com/aliencaocao/vlm-for-memes-aisg)


## Training Details

### Training Data

The model was trained on the [multimodal_meme_classification_singapore](https://huggingface.co/datasets/aliencaocao/multimodal_meme_classification_singapore) dataset. This dataset contains memes labeled as offensive or not within the Singaporean context.

### Training Procedure

More details about the training procedure can be found in the paper.

## Evaluation

The model achieved an AUROC of 0.8192 and an accuracy of 0.8043 on a held-out test set.  See the paper for more details on the evaluation methodology.

## Citation

```bibtex
@misc{yuxuan2025detectingoffensivememessocial,
      title={Detecting Offensive Memes with Social Biases in Singapore Context Using Multimodal Large Language Models}, 
      author={Cao Yuxuan and Wu Jiayang and Alistair Cheong Liang Chuen and Bryan Shan Guanrong and Theodore Lee Chong Jen and Sherman Chann Zhi Shen},
      year={2025},
      eprint={2502.18101},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2502.18101}, 
}
```