|
--- |
|
dataset_info: |
|
features: |
|
- name: text |
|
dtype: string |
|
- name: user_age |
|
dtype: int64 |
|
- name: user_gender |
|
dtype: string |
|
- name: text_topic |
|
dtype: string |
|
- name: class |
|
dtype: string |
|
- name: age |
|
dtype: int64 |
|
- name: text_topic_eng |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 751331 |
|
num_examples: 3770 |
|
download_size: 254089 |
|
dataset_size: 751331 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
license: cc-by-sa-4.0 |
|
language: |
|
- ko |
|
tags: |
|
- safety |
|
--- |
|
|
|
reference: [https://github.com/jason9693/APEACH](https://github.com/jason9693/APEACH) |
|
``` |
|
@inproceedings{yang-etal-2022-apeach, |
|
title = "{APEACH}: Attacking Pejorative Expressions with Analysis on Crowd-Generated Hate Speech Evaluation Datasets", |
|
author = "Yang, Kichang and |
|
Jang, Wonjun and |
|
Cho, Won Ik", |
|
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2022", |
|
month = dec, |
|
year = "2022", |
|
address = "Abu Dhabi, United Arab Emirates", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://aclanthology.org/2022.findings-emnlp.525", |
|
pages = "7076--7086", |
|
abstract = "In hate speech detection, developing training and evaluation datasets across various domains is the critical issue. Whereas, major approaches crawl social media texts and hire crowd-workers to annotate the data. Following this convention often restricts the scope of pejorative expressions to a single domain lacking generalization. Sometimes domain overlap between training corpus and evaluation set overestimate the prediction performance when pretraining language models on low-data language. To alleviate these problems in Korean, we propose APEACH that asks unspecified users to generate hate speech examples followed by minimal post-labeling. We find that APEACH can collect useful datasets that are less sensitive to the lexical overlaps between the pretraining corpus and the evaluation set, thereby properly measuring the model performance.", |
|
} |
|
``` |