sanjay73 commited on
Commit
e837b0e
·
verified ·
1 Parent(s): bffe3af

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -325,17 +325,17 @@ task_categories:
325
  </a>
326
  </div>
327
 
328
- **Pralekha** is a large-scale parallel document dataset for **Cross-Lingual Document Alignment (CLDA)** and **Machine Translation (MT)** across **11 Indic languages** and **English**. It comprises over **3 million** document pairs, with **1.5 million** being English-centric.
329
 
330
  ---
331
 
332
  ## Dataset Description
333
 
334
- **Pralekha** covers 12 languages—Bengali (`ben`), Gujarati (`guj`), Hindi (`hin`), Kannada (`kan`), Malayalam (`mal`), Marathi (`mar`), Odia (`ori`), Punjabi (`pan`), Tamil (`tam`), Telugu (`tel`), Urdu (`urd`), and English (`eng`). It includes a mixture of high- and medium-resource languages, covering 11 different scripts. The dataset spans two broad domains: **News Bulletins** ([Indian Press Information Bureau (PIB)](https://pib.gov.in)) and **Podcast Scripts** ([Mann Ki Baat](https://www.pmindia.gov.in/en/mann-ki-baat)), offering both written and spoken forms of data. All the data is human-written or human-verified, ensuring high quality.
335
 
336
  While this accounts for `alignable` (parallel) documents, In real-world scenarios, multilingual corpora often include `unalignable` documents. To simulate this for CLDA evaluation, we sample `unalignable` documents from [Sangraha Unverified](https://huggingface.co/datasets/ai4bharat/sangraha/viewer/unverified), selecting 50% of Pralekha’s size to maintain a 1:2 ratio of `unalignable` to `alignable` documents.
337
 
338
- For Machine Translation (MT) tasks, we first randomly sample 1,000 documents per English-Indic language pair, ensuring a good distribution of varying document lengths. After excluding these sampled documents, we use the remaining documents for training document-level machine translation models.
339
 
340
  ---
341
 
@@ -352,7 +352,7 @@ For Machine Translation (MT) tasks, we first randomly sample 1,000 documents per
352
 
353
  - **`src_lang`:** Source Language (eng)
354
  - **`src_text`:** Source Language Text
355
- - **`tgt_lang`:** Target Language ((ISO 639-3 code)
356
  - **`tgt_text`:** Target Language Text
357
 
358
  ---
 
325
  </a>
326
  </div>
327
 
328
+ **Pralekha** is a large-scale parallel document dataset spanning across **11 Indic languages** and **English**. It comprises over **3 million** document pairs, with **1.5 million** being English-centric. This dataset serves both as a benchmark for evaluating **Cross-Lingual Document Alignment (CLDA)** techniques and as a domain-specific parallel corpus for training document-level **Machine Translation (MT)** models in Indic Languages.
329
 
330
  ---
331
 
332
  ## Dataset Description
333
 
334
+ **Pralekha** covers 12 languages—Bengali (`ben`), Gujarati (`guj`), Hindi (`hin`), Kannada (`kan`), Malayalam (`mal`), Marathi (`mar`), Odia (`ori`), Punjabi (`pan`), Tamil (`tam`), Telugu (`tel`), Urdu (`urd`), and English (`eng`). It includes a mixture of high- and medium-resource languages, covering 11 different scripts. The dataset spans two broad domains: **News Bulletins** ([Indian Press Information Bureau (PIB)](https://pib.gov.in)) and **Podcast Scripts** ([Mann Ki Baat (MKB)](https://www.pmindia.gov.in/en/mann-ki-baat)), offering both written and spoken forms of data. All the data is human-written or human-verified, ensuring high quality.
335
 
336
  While this accounts for `alignable` (parallel) documents, In real-world scenarios, multilingual corpora often include `unalignable` documents. To simulate this for CLDA evaluation, we sample `unalignable` documents from [Sangraha Unverified](https://huggingface.co/datasets/ai4bharat/sangraha/viewer/unverified), selecting 50% of Pralekha’s size to maintain a 1:2 ratio of `unalignable` to `alignable` documents.
337
 
338
+ For Machine Translation (MT) tasks, we first randomly sample 1,000 documents from the `alignable` subset per English-Indic language pair for each development (dev) and test set, ensuring a good distribution of varying document lengths. After excluding these sampled documents, we use the remaining documents as the training set for training document-level machine translation models.
339
 
340
  ---
341
 
 
352
 
353
  - **`src_lang`:** Source Language (eng)
354
  - **`src_text`:** Source Language Text
355
+ - **`tgt_lang`:** Target Language (ISO 639-3 code)
356
  - **`tgt_text`:** Target Language Text
357
 
358
  ---