---
license: mit
dataset_info:
- config_name: 100_tos
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 5240826
num_examples: 92
download_size: 2497746
dataset_size: 5240826
- config_name: 10_tos
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 1920213
num_examples: 20
download_size: 718890
dataset_size: 1920213
- config_name: 142_tos
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 12968483
num_examples: 140
download_size: 4884205
dataset_size: 12968483
- config_name: cuad
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 1180620
num_examples: 28
download_size: 484787
dataset_size: 1180620
- config_name: memnet_tos
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 5607746
num_examples: 100
download_size: 2012157
dataset_size: 5607746
- config_name: multilingual_unfair_clause
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 22775210
num_examples: 200
download_size: 9557263
dataset_size: 22775210
- config_name: polisis
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 3137858
num_examples: 4570
- name: validation
num_bytes: 802441
num_examples: 1153
- name: test
num_bytes: 967678
num_examples: 1446
download_size: 1827549
dataset_size: 4907977
- config_name: privacy_glue__piextract
features:
- name: document
dtype: string
splits:
- name: validation
num_bytes: 7106934
num_examples: 4116
- name: train
num_bytes: 18497078
num_examples: 12140
download_size: 5707087
dataset_size: 25604012
- config_name: privacy_glue__policy_detection
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 13657226
num_examples: 1301
download_size: 6937382
dataset_size: 13657226
- config_name: privacy_glue__policy_ie
features:
- name: type_i
dtype: string
- name: type_ii
dtype: string
splits:
- name: test
num_bytes: 645788
num_examples: 6
- name: train
num_bytes: 2707213
num_examples: 25
download_size: 1097051
dataset_size: 3353001
- config_name: privacy_glue__policy_qa
features:
- name: document
dtype: string
splits:
- name: test
num_bytes: 1353787
num_examples: 20
- name: dev
num_bytes: 1230490
num_examples: 20
- name: train
num_bytes: 5441319
num_examples: 75
download_size: 2418472
dataset_size: 8025596
- config_name: privacy_glue__polisis
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 3073878
num_examples: 4570
- name: validation
num_bytes: 786299
num_examples: 1153
- name: test
num_bytes: 947434
num_examples: 1446
download_size: 1816140
dataset_size: 4807611
- config_name: privacy_glue__privacy_qa
features:
- name: document
dtype: string
splits:
- name: train
num_bytes: 12099109
num_examples: 27
- name: test
num_bytes: 4468753
num_examples: 8
download_size: 1221943
dataset_size: 16567862
configs:
- config_name: 100_tos
data_files:
- split: train
path: 100_tos/train-*
- config_name: 10_tos
data_files:
- split: train
path: 10_tos/train-*
- config_name: 142_tos
data_files:
- split: train
path: 142_tos/train-*
- config_name: cuad
data_files:
- split: train
path: cuad/train-*
- config_name: memnet_tos
data_files:
- split: train
path: memnet_tos/train-*
- config_name: multilingual_unfair_clause
data_files:
- split: train
path: multilingual_unfair_clause/train-*
- config_name: polisis
data_files:
- split: train
path: privacy_glue/polisis/train-*
- split: validation
path: privacy_glue/polisis/validation-*
- split: test
path: privacy_glue/polisis/test-*
- config_name: privacy_glue__piextract
data_files:
- split: validation
path: privacy_glue/piextract/validation-*
- split: train
path: privacy_glue/piextract/train-*
- config_name: privacy_glue__policy_detection
data_files:
- split: train
path: privacy_glue/policy_detection/train-*
- config_name: privacy_glue__policy_ie
data_files:
- split: test
path: privacy_glue/policy_ie/test-*
- split: train
path: privacy_glue/policy_ie/train-*
- config_name: privacy_glue__policy_qa
data_files:
- split: test
path: privacy_glue/policy_qa/test-*
- split: dev
path: privacy_glue/policy_qa/dev-*
- split: train
path: privacy_glue/policy_qa/train-*
- config_name: privacy_glue__polisis
data_files:
- split: train
path: privacy_glue/polisis/train-*
- split: validation
path: privacy_glue/polisis/validation-*
- split: test
path: privacy_glue/polisis/test-*
- config_name: privacy_glue__privacy_qa
data_files:
- split: train
path: privacy_glue/privacy_qa/train-*
- split: test
path: privacy_glue/privacy_qa/test-*
---
# A collection of Terms of Service or Privacy Policy datasets
## Annotated datasets
### CUAD
Specifically, the 28 service agreements from [CUAD](https://www.atticusprojectai.org/cuad), which are licensed under CC BY 4.0 (subset: `cuad`).
Code
```python
import datasets
from tos_datasets.proto import DocumentQA
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "cuad")
print(DocumentQA.model_validate_json(ds["document"][0]))
```
### 100 ToS
From [Annotated 100 ToS](https://data.mendeley.com/datasets/dtbj87j937/3), CC BY 4.0 (subset: `100_tos`).
Code
```python
import datasets
from tos_datasets.proto import DocumentEUConsumerLawAnnotation
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "100_tos")
print(DocumentEUConsumerLawAnnotation.model_validate_json(ds["document"][0]))
```
### Multilingual Unfair Clause
From [CLAUDETTE](http://claudette.eui.eu/corpora/index.html)/[Multilingual Unfair Clause](https://github.com/nlp-unibo/Multilingual-Unfair-Clause-Detection), CC BY 4.0 (subset: `multilingual_unfair_clause`).
It was built from [CLAUDETTE](http://claudette.eui.eu/corpora/index.html)/[25 Terms of Service in English, Italian, German, and Polish (100 documents in total) from A Corpus for Multilingual Analysis of Online Terms of Service](http://claudette.eui.eu/corpus_multilingual_NLLP2021.zip).
Code
```python
import datasets
from tos_datasets.proto import DocumentClassification
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "multilingual_unfair_clause")
print(DocumentClassification.model_validate_json(ds["document"][0]))
```
### Memnet ToS
From [100 Terms of Service in English from Detecting and explaining unfairness in consumer contracts through memory networks](https://github.com/federicoruggeri/Memnet_ToS), MIT (subset: `memnet_tos`).
Code
```python
import datasets
from tos_datasets.proto import DocumentClassification
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "memnet_tos")
print(DocumentClassification.model_validate_json(ds["document"][0]))
```
### 142 ToS
From [142 Terms of Service in English divided according to market sector from Assessing the Cross-Market Generalization Capability of the CLAUDETTE System](http://claudette.eui.eu/corpus_142_ToS.zip), Unknown (subset: `142_tos`). This should also includes [50 Terms of Service in English from "CLAUDETTE: an Automated Detector of Potentially Unfair Clauses in Online Terms of Service"](http://claudette.eui.eu/ToS.zip).
Code
```python
import datasets
from tos_datasets.proto import DocumentClassification
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "142_tos")
print(DocumentClassification.model_validate_json(ds["document"][0]))
```
### 10 ToS/PP
From [5 Terms of Service and 5 Privacy Policies in English and German (10 documents in total) from Cross-lingual Annotation Projection in Legal Texts](https://bitbucket.org/a-galaxy/cross-lingual-annotation-projection-in-legal-texts), GNU GPL 3.0 (subset: `10_tos`)
Code
```python
import datasets
from tos_datasets.proto import DocumentClassification
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "10_tos")
print(DocumentClassification.model_validate_json(ds["document"][0]))
```
### PolicyQA
> [!IMPORTANT]
> This dataset seems to have some annotation issues where __unanswerable__ questions are still answered with SQuAD-v1 format instead of the v2 format.
From [PolicyQA](https://github.com/wasiahmad/PolicyQA), MIT (subset: `privacy_glue/policy_qa`).
Code
```python
import datasets
from tos_datasets.proto import DocumentQA
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "privacy_glue/policy_qa")
print(DocumentQA.model_validate_json(ds["train"]["document"][0]))
```
### PolicyIE
From [PolicyIE](https://github.com/wasiahmad/PolicyIE), MIT (subset: `privacy_glue/policy_ie`).
Code
```python
import datasets
from tos_datasets.proto import DocumentSequenceClassification, DocumentEvent
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "privacy_glue/policy_ie")
print(DocumentSequenceClassification.model_validate_json(ds["train"]["type_i"][0]))
print(DocumentEvent.model_validate_json(ds["train"]["type_ii"][0]))
```
### Policy Detection
From [policy-detection-data](, GPL 3.0 (subset: `privacy_glue/policy_detection`).
Code
```python
import datasets
from tos_datasets.proto import DocumentClassification
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "privacy_glue/policy_detection")
print(DocumentClassification.model_validate_json(ds["train"]["document"][0]))
```
### Polisis
From [Polisis](https://github.com/SmartDataAnalytics/Polisis_Benchmark), Unknown (subset: `privacy_glue/polisis`).
Code
```python
import datasets
from tos_datasets.proto import DocumentClassification
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "privacy_glue/polisis")
print(DocumentClassification.model_validate_json(ds["test"]["document"][0]))
```
### PrivacyQA
From [PrivacyQA](https://github.com/AbhilashaRavichander/PrivacyQA_EMNLP), MIT (subset: `privacy_qa`).
Code
```python
import datasets
from tos_datasets.proto import DocumentClassification
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "privacy_glue/privacy_qa")
print(DocumentClassification.model_validate_json(ds["test"]["document"][0]))
```
### Piextract
From [Piextract](https://github.com/um-rtcl/piextract_dataset), Unknown (subset: `privacy_glue/piextract`).
Code
```python
import datasets
from tos_datasets.proto import DocumentSequenceClassification
ds = datasets.load_dataset("chenghao/tos_pp_dataset", "privacy_glue/piextract")
print(DocumentSequenceClassification.model_validate_json(ds["train"]["document"][0]))
```
## WIP
- [Annotated Italian TOS sentences](https://github.com/i3-fbk/LLM-PE_Terms_and_Conditions_Contracts), Apache 2.0 Only sentence level annotations, missing original full text
- [Huggingface](https://huggingface.co/datasets/CodeHima/TOS_Dataset), MIT Only sentence level annotations, missing original full text
- [ ] [ToSDR API](https://developers.tosdr.org/dev/get-service-v2), Unknown