Datasets:

Modalities:
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
Dask
File size: 1,842 Bytes
35f5b4e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
62d1515
35f5b4e
62d1515
 
 
 
 
 
 
 
bcde28e
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
---
dataset_info:
  features:
  - name: meta
    struct:
    - name: arxiv_id
      dtype: string
    - name: language
      dtype: string
    - name: timestamp
      dtype: string
    - name: url
      dtype: string
    - name: yymm
      dtype: string
  - name: text
    dtype: string
  splits:
  - name: train
    num_bytes: 857168232
    num_examples: 13155
  download_size: 382068275
  dataset_size: 857168232
---
# ArXiv papers from RedPajama-Data originally published in February 2023

We collect the ArXiv papers released shortly before the training data cutoff date for the [OpenLLaMA models](https://huggingface.co/openlm-research/open_llama_7b). 

The OpenLLaMA models (V1) have been trained on [RedPajama data](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T).
The last batch of ArXiv papers included in this dataset are papers published in February 2023. 
To get the members close to the cutoff data, we collect the 13,155 papers published in "2302" as part of the training dataset. 
We process the raw LateX files using this [script](https://github.com/togethercomputer/RedPajama-Data/blob/rp_v1/data_prep/arxiv/run_clean.py). 

This dataset has been used as source for 'member' documents to develop (document-level) MIAs against LLMs using data collected shortly before (member) and after (non-member) the training cutoff date for the target model ([the suite of OpenLLaMA models](https://huggingface.co/openlm-research/open_llama_7b)). 
For non-members for the RDD setup, we refer to our [Github repo](https://github.com/computationalprivacy/mia_llms_benchmark/tree/main/document_level). 
For more details and results see the section of Regression Discontiuity Design (RDD) in the paper ["SoK: Membership Inference Attacks on LLMs are Rushing Nowhere (and How to Fix It)"](https://arxiv.org/pdf/2406.17975).