Hyukkyu commited on
Commit
371272d
·
verified ·
1 Parent(s): f676491

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +261 -14
README.md CHANGED
@@ -320,32 +320,279 @@ configs:
320
  ---
321
  # SuperGLUE Benchmark Datasets
322
 
323
- This repository contains the [**SuperGLUE**](https://arxiv.org/pdf/1905.00537) benchmark datasets uploaded to the Hugging Face Hub. Each dataset is available as a separate configuration, making it easy to load individual datasets using the [datasets](https://github.com/huggingface/datasets) library.
324
 
325
- ## Datasets Included
326
 
327
- The repository includes the following SuperGLUE datasets:
328
 
329
- - **BoolQ**
330
- - **CB**
331
- - **COPA**
332
- - **MultiRC**
333
- - **ReCoRD**
334
- - **RTE**
335
- - **WiC**
336
- - **WSC**
 
 
 
 
 
 
 
337
 
338
  Each dataset has been preprocessed to ensure consistency across train, validation, and test splits. Missing keys in the test split have been filled with dummy values (type-aware) to match the features found in the training and validation splits.
339
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
340
  ## Usage
341
 
342
- You can load any of the datasets using the Hugging Face `datasets` library. For example, to load the BoolQ dataset, run:
343
 
344
  ```python
345
  from datasets import load_dataset
346
 
347
  # Load the BoolQ dataset from the SuperGLUE benchmark
348
- dataset = load_dataset("Hyukkyu/superglue", "BoolQ")
349
 
350
  # Access train, validation, and test splits
351
  train_split = dataset["train"]
@@ -356,7 +603,7 @@ print(train_split)
356
  ```
357
  Replace "BoolQ" with the desired configuration name (e.g., "CB", "COPA", "MultiRC", etc.) to load other datasets.
358
 
359
- ## Data Processing
360
  - Schema Consistency:
361
  A recursive procedure was used to infer the schema from the train and validation splits and fill in missing keys in the test split with dummy values. This ensures that all splits have the same features, preventing issues during model training or evaluation.
362
  - Type-Aware Dummy Values:
 
320
  ---
321
  # SuperGLUE Benchmark Datasets
322
 
323
+ This repository contains the [**SuperGLUE**](https://arxiv.org/pdf/1905.00537) benchmark datasets. Each dataset is available as a separate configuration, making it easy to load individual datasets using the [datasets](https://github.com/huggingface/datasets) library.
324
 
325
+ ## Dataset Descriptions
326
 
327
+ ### Datasets Included
328
 
329
+ - **BoolQ:** A question-answering task where each example consists of a short passage and a yes/no question about the passage. The questions are provided anonymously and unsolicited by users of the Google search engine and paired with a paragraph from a Wikipedia article containing the answer.
330
+
331
+ - **CB (CommitmentBank):** A natural language inference task where each example consists of a premise containing an embedded clause, and the corresponding hypothesis is the extraction of that clause. The task focuses on determining whether the premise entails the hypothesis, contradicts it, or is neutral.
332
+
333
+ - **COPA (Choice of Plausible Alternatives):** A causal reasoning task where the system selects the more plausible alternative between two choices given a premise. Each example consists of a premise and two possible alternatives, and the task is to choose the alternative that has a causal relationship with the premise.
334
+
335
+ - **MultiRC (Multiple Sentence Reading Comprehension):** A reading comprehension task where each example consists of a context paragraph, a question about the paragraph, and a list of possible answers. The task requires identifying all correct answers for each question, and there may be multiple correct answers.
336
+
337
+ - **ReCoRD (Reading Comprehension with Commonsense Reasoning Dataset):** A cloze-style reading comprehension task that evaluates a model’s ability to use commonsense reasoning to predict which entity is missing from a passage. Each example consists of a passage with a missing entity and a list of possible entities to fill in the blank.
338
+
339
+ - **RTE (Recognizing Textual Entailment):** A textual entailment task that involves determining whether a given premise entails a hypothesis. Each example consists of a premise and a hypothesis, and the task is to predict whether the hypothesis is true based on the premise.
340
+
341
+ - **WiC (Word-in-Context):** A word sense disambiguation task that determines if a word is used in the same sense in two different contexts. Each example consists of two sentences containing the same word, and the task is to decide whether the word has the same meaning in both sentences.
342
+
343
+ - **WSC (Winograd Schema Challenge):** A pronoun resolution task where the system must determine the antecedent of a pronoun in a sentence. Each example consists of a sentence with a pronoun and a list of possible antecedents, and the task is to select the correct antecedent.
344
 
345
  Each dataset has been preprocessed to ensure consistency across train, validation, and test splits. Missing keys in the test split have been filled with dummy values (type-aware) to match the features found in the training and validation splits.
346
 
347
+ ### Languages
348
+
349
+ All data in SuperGLUE is in English.
350
+
351
+ ## Dataset Structure
352
+
353
+ ### Data Instances
354
+
355
+ Each task in SuperGLUE is split into train, validation, and test sets. An example instance from each configuration is provided in the task-specific sections below.
356
+
357
+ #### BoolQ
358
+ - train split: 9427 examples
359
+ - validation split: 3270 examples
360
+ - test split: 3245 examples
361
+
362
+ An example of 'train' looks as follows.
363
+ ```json
364
+ {
365
+ "question": "do iran and afghanistan speak the same language",
366
+ "passage": "Persian language -- Persian (/\u02c8p\u025c\u02d0r\u0292\u0259n, -\u0283\u0259n/), also known by its endonym Farsi (\u0641\u0627\u0631\u0633\u06cc f\u0101rsi (f\u0252\u02d0\u027e\u02c8si\u02d0) ( listen)), is one of the Western Iranian languages within the Indo-Iranian branch of the Indo-European language family. It is primarily spoken in Iran, Afghanistan (officially known as Dari since 1958), and Tajikistan (officially known as Tajiki since the Soviet era), and some other regions which historically were Persianate societies and considered part of Greater Iran. It is written in the Persian alphabet, a modified variant of the Arabic script, which itself evolved from the Aramaic alphabet.",
367
+ "idx": 0,
368
+ "label": true
369
+ }
370
+ ```
371
+
372
+ #### CB
373
+ - train split: 250 examples
374
+ - validation split: 56 examples
375
+ - test split: 250 examples
376
+
377
+ An example of 'train' looks as follows.
378
+ ```json
379
+ {
380
+ "premise": "It was a complex language. Not written down but handed down. One might say it was peeled down.",
381
+ "hypothesis": "the language was peeled down",
382
+ "label": "entailment",
383
+ "idx": 0
384
+ }
385
+ ```
386
+
387
+ #### COPA
388
+ - train split: 400 examples
389
+ - validation split: 100 examples
390
+ - test split: 500 examples
391
+
392
+ An example of 'train' looks as follows.
393
+ ```json
394
+ {
395
+ "premise": "My body cast a shadow over the grass.",
396
+ "choice1": "The sun was rising.",
397
+ "choice2": "The grass was cut.",
398
+ "question": "cause",
399
+ "label": 0,
400
+ "idx": 0
401
+ }
402
+ ```
403
+ #### MultiRC
404
+ - train split: 456 examples
405
+ - validation split: 83 examples
406
+ - test split: 166 examples
407
+
408
+ An example of 'train' looks as follows.
409
+ ```json
410
+ {
411
+ "idx": 0,
412
+ "version": 1.1,
413
+ "passage": {
414
+ "questions": [
415
+ {
416
+ "answers": [
417
+ {
418
+ "idx": 0,
419
+ "label": 0,
420
+ "text": "Children, Gerd, or Dorian Popa"
421
+ },
422
+ {
423
+ "idx": 1,
424
+ "label": 0,
425
+ "text": "Monetary rewards"
426
+ },
427
+ {
428
+ "idx": 2,
429
+ "label": 1,
430
+ "text": "Asking Pakistan to help the USA"
431
+ },
432
+ {
433
+ "idx": 3,
434
+ "label": 1,
435
+ "text": "Meeting with General Musharraf"
436
+ },
437
+ {
438
+ "idx": 4,
439
+ "label": 1,
440
+ "text": "President Clinton offered the moon in terms of better relations with the United States"
441
+ },
442
+ {
443
+ "idx": 5,
444
+ "label": 0,
445
+ "text": "A Presidential visit in March"
446
+ },
447
+ {
448
+ "idx": 6,
449
+ "label": 1,
450
+ "text": "Paper checks"
451
+ },
452
+ {
453
+ "idx": 7,
454
+ "label": 0,
455
+ "text": "Increasing trade with Pakistan"
456
+ },
457
+ {
458
+ "idx": 8,
459
+ "label": 1,
460
+ "text": "Persuading Pakistan to use its influence over the Taliban by dangling before him the possibility of a presidential visit in March as a reward for Pakistani cooperation"
461
+ }
462
+ ],
463
+ "idx": 0,
464
+ "question": "What did the high-level effort to persuade Pakistan include?"
465
+ },
466
+ ...
467
+ ],
468
+ "text": "While this process moved along, diplomacy ..."
469
+ }
470
+ ```
471
+
472
+ #### ReCoRD
473
+ - train split: 65709 examples
474
+ - validation split: 7481 examples
475
+ - test split: 7484 examples
476
+
477
+ An example of 'train' looks as follows.
478
+ ```json
479
+
480
+ "source": "Daily mail",
481
+ "passage": {
482
+ "entities": [
483
+ {
484
+ "end": 96,
485
+ "start": 86
486
+ },
487
+ {
488
+ "end": 183,
489
+ "start": 178
490
+ },
491
+ {
492
+ "end": 206,
493
+ "start": 197
494
+ },
495
+ {
496
+ "end": 361,
497
+ "start": 357
498
+ },
499
+ {
500
+ "end": 539,
501
+ "start": 535
502
+ },
503
+ {
504
+ "end": 636,
505
+ "start": 627
506
+ },
507
+ {
508
+ "end": 681,
509
+ "start": 672
510
+ }
511
+ ],
512
+ "text": "The harrowing stories of women and children locked up for so-called 'moral crimes' in Afghanistan's notorious female prison have been revealed after cameras were allowed inside. Mariam has been in Badam Bagh prison for three months after she shot a man who just raped her at gunpoint and then turned the weapon on herself - but she has yet to been charged. Nuria has eight months left to serve of her sentence for trying to divorce her husband. She gave birth in prison to her son and they share a cell together. Scroll down for video Nuria was jailed for trying to divorce her husband. Her son is one of 62 children living at Badam Bagh prison\n@highlight\nMost of the 202 Badam Bagh inmates are jailed for so-called 'moral crimes'\n@highlight\nCrimes include leaving their husbands or refusing an arrange marriage\n@highlight\n62 children live there and share cells with their mothers and five others"
513
+ },
514
+ "qas": [
515
+ {
516
+ "answers": [
517
+ {
518
+ "end": 539,
519
+ "start": 535,
520
+ "text": "Nuria"
521
+ }
522
+ ],
523
+ "idx": 0,
524
+ "query": "The baby she gave birth to is her husbands and he has even offered to have the courts set her free if she returns, but @placeholder has refused."
525
+ }
526
+ ],
527
+ "idx": 0
528
+ }
529
+ ```
530
+
531
+ #### RTE
532
+ - train split: 2490 examples
533
+ - validation split: 277 examples
534
+ - test split: 3000 examples
535
+
536
+ An example of 'train' looks as follows.
537
+ ```json
538
+ {
539
+ "premise": "No Weapons of Mass Destruction Found in Iraq Yet.",
540
+ "hypothesis": "Weapons of Mass Destruction Found in Iraq.",
541
+ "label": "not_entailment",
542
+ "idx": 0
543
+ }
544
+ ```
545
+
546
+ #### WiC
547
+ - train split: 5428 examples
548
+ - validation split: 638 examples
549
+ - test split: 1400 examples
550
+
551
+ An example of 'train' looks as follows.
552
+ ```json
553
+ {
554
+ "word": "place",
555
+ "sentence1": "Do you want to come over to my place later?",
556
+ "sentence2": "A political system with no place for the less prominent groups.",
557
+ "idx": 0,
558
+ "label": false,
559
+ "start1": 31,
560
+ "start2": 27,
561
+ "end1": 36,
562
+ "end2": 32,
563
+ "version": 1.1
564
+ }
565
+ ```
566
+
567
+ #### WSC
568
+ - train split: 554 examples
569
+ - validation split: 104 examples
570
+ - test split: 146 examples
571
+
572
+ An example of 'train' looks as follows.
573
+ ```json
574
+ {
575
+ "text": "Mark told Pete many lies about himself, which Pete included in his book. He should have been more skeptical.",
576
+ "target": {
577
+ "span1_index": 0,
578
+ "span1_text": "Mark",
579
+ "span2_index": 13,
580
+ "span2_text": "He"
581
+ },
582
+ "idx": 0,
583
+ "label": false
584
+ }
585
+ ```
586
+
587
  ## Usage
588
 
589
+ You can load any of the datasets using the Huggingface `datasets` library. For example, to load the BoolQ dataset, run:
590
 
591
  ```python
592
  from datasets import load_dataset
593
 
594
  # Load the BoolQ dataset from the SuperGLUE benchmark
595
+ dataset = load_dataset("Hyukkyu/superglue", "boolq")
596
 
597
  # Access train, validation, and test splits
598
  train_split = dataset["train"]
 
603
  ```
604
  Replace "BoolQ" with the desired configuration name (e.g., "CB", "COPA", "MultiRC", etc.) to load other datasets.
605
 
606
+ ## Data Preprocessing
607
  - Schema Consistency:
608
  A recursive procedure was used to infer the schema from the train and validation splits and fill in missing keys in the test split with dummy values. This ensures that all splits have the same features, preventing issues during model training or evaluation.
609
  - Type-Aware Dummy Values: