Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ language:
|
|
16 |
- tt
|
17 |
license: openrail++
|
18 |
size_categories:
|
19 |
-
-
|
20 |
task_categories:
|
21 |
- text-generation
|
22 |
dataset_info:
|
@@ -76,14 +76,15 @@ configs:
|
|
76 |
path: data/ar-*
|
77 |
- split: hi
|
78 |
path: data/hi-*
|
|
|
|
|
79 |
---
|
80 |
**Multilingual Text Detoxification with Parallel Data**
|
81 |
|
82 |
[](https://aclanthology.org/2025.coling-main.535/)
|
83 |
[](https://ceur-ws.org/Vol-3740/paper-223.pdf)
|
84 |
|
85 |
-
This is the multilingual parallel dataset for the text detoxification task
|
86 |
-
For each of 9 languages, we collected 1k pairs of toxic<->detoxified instances splitted into two parts: dev (400 pairs) and test (600 pairs).
|
87 |
|
88 |
📰 **Updates**
|
89 |
|
@@ -99,6 +100,9 @@ For each of 9 languages, we collected 1k pairs of toxic<->detoxified instances s
|
|
99 |
|
100 |
**[2022]** You can also check previously created training corpora: [English ParaDetox](https://huggingface.co/datasets/s-nlp/paradetox) from ACL 2022 and [Russian ParaDetox](https://huggingface.co/datasets/s-nlp/ru_paradetox).
|
101 |
|
|
|
|
|
|
|
102 |
## Toxic Samples Sources
|
103 |
|
104 |
The list of the sources for the original toxic sentences:
|
@@ -212,4 +216,8 @@ If you would like to acknowledge our work, please, cite the following manuscript
|
|
212 |
biburl = {https://dblp.org/rec/conf/ecir/BevendorffCCDEFFKMMPPRRSSSTUWZ24.bib},
|
213 |
bibsource = {dblp computer science bibliography, https://dblp.org}
|
214 |
}
|
215 |
-
```
|
|
|
|
|
|
|
|
|
|
16 |
- tt
|
17 |
license: openrail++
|
18 |
size_categories:
|
19 |
+
- 10K<n<100K
|
20 |
task_categories:
|
21 |
- text-generation
|
22 |
dataset_info:
|
|
|
76 |
path: data/ar-*
|
77 |
- split: hi
|
78 |
path: data/hi-*
|
79 |
+
tags:
|
80 |
+
- toxic
|
81 |
---
|
82 |
**Multilingual Text Detoxification with Parallel Data**
|
83 |
|
84 |
[](https://aclanthology.org/2025.coling-main.535/)
|
85 |
[](https://ceur-ws.org/Vol-3740/paper-223.pdf)
|
86 |
|
87 |
+
This is the multilingual parallel dataset for the text detoxification task. Prepared for TextDetox Shared Task.
|
|
|
88 |
|
89 |
📰 **Updates**
|
90 |
|
|
|
100 |
|
101 |
**[2022]** You can also check previously created training corpora: [English ParaDetox](https://huggingface.co/datasets/s-nlp/paradetox) from ACL 2022 and [Russian ParaDetox](https://huggingface.co/datasets/s-nlp/ru_paradetox).
|
102 |
|
103 |
+
**[2022]** The first release for [CLEF TextDetox 2024](https://pan.webis.de/clef24/pan24-web/text-detoxification.html) shared task.
|
104 |
+
For each of 9 languages, we collected 1k pairs of toxic<->detoxified instances splitted into two parts: dev (400 pairs) and test (600 pairs).
|
105 |
+
|
106 |
## Toxic Samples Sources
|
107 |
|
108 |
The list of the sources for the original toxic sentences:
|
|
|
216 |
biburl = {https://dblp.org/rec/conf/ecir/BevendorffCCDEFFKMMPPRRSSSTUWZ24.bib},
|
217 |
bibsource = {dblp computer science bibliography, https://dblp.org}
|
218 |
}
|
219 |
+
```
|
220 |
+
|
221 |
+
# Corresponding Contact
|
222 |
+
|
223 |
+
[Daryna Dementieva](https://huggingface.co/dardem)
|