dv-synthetic-errors / README.md
alakxender's picture
Update README.md
ab5312a verified
---
dataset_info:
features:
- name: correct
dtype: string
- name: incorrect
dtype: string
splits:
- name: train
num_bytes: 1211373359.8361242
num_examples: 3161164
- name: test
num_bytes: 151421861.5819379
num_examples: 395146
- name: validation
num_bytes: 151421861.5819379
num_examples: 395146
download_size: 752362217
dataset_size: 1514217083
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: validation
path: data/validation-*
language:
- dv
license: apache-2.0
pretty_name: dv_text_erros
---
# DV Text Errors
Dhivehi text error correction dataset containing correct sentences and synthetically generated errors. The dataset aims to test Dhivehi language error correction models and tools.
## About Dataset
- **Task**: Text error correction
- **Language**: Dhivehi (dv)
## Dataset Structure
Input-output pairs of Dhivehi text:
- `correct`: Original correct sentences
- `incorrect`: Sentences with synthetic errors
## Statistics
- Train set: {train_examples} examples ({0.7999997975429817}%)
- Test set: {test_examples} examples ({0.10000010122850919}%)
- Validation set: {val_examples} examples ({0.10000010122850919}%)
**Details:**
- Unique words: {448628}
```json
{
"total_examples": {
"train": 3161164,
"test": 395146,
"validation": 395146
},
"avg_sentence_length": {
"train": 11.968980097204701,
"test": 11.961302910822836,
"validation": 11.973824864733542
},
"error_distribution": {
"min": 0,
"max": 2411,
"avg": 64.85144965588626
}
}
```
## Usage
```python
from datasets import load_dataset
dataset = load_dataset("alakxender/dv-synthetic-errors")
```
## Dataset Creation
Created using:
- Source: Collection of Dhivehi articles
- Error generation: Character and diacritic substitutions
- Error rate: 30% per word probability