crrodrvi commited on
Commit
3ca4048
·
verified ·
1 Parent(s): bba53d9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -15
README.md CHANGED
@@ -1,20 +1,25 @@
1
  ---
2
  license: mit
3
- base_model: facebook/bart-large-cnn
4
  tags:
5
- - text2text-generation
6
  - generated_from_trainer
 
 
7
  model-index:
8
- - name: modelo_simplificacion_bart
9
  results: []
 
 
 
 
 
10
  ---
11
-
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
-
15
  # modelo_simplificacion_bart
16
 
17
- This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the None dataset.
 
 
 
18
 
19
  ## Model description
20
 
@@ -33,24 +38,27 @@ More information needed
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
- - learning_rate: 5e-05
37
  - train_batch_size: 4
38
  - eval_batch_size: 4
39
  - seed: 42
40
- - gradient_accumulation_steps: 16
41
- - total_train_batch_size: 64
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
- - lr_scheduler_warmup_steps: 500
45
- - num_epochs: 1
46
 
47
  ### Training results
48
 
 
 
 
 
 
 
49
 
50
 
51
  ### Framework versions
52
 
53
- - Transformers 4.42.4
54
- - Pytorch 2.3.1+cu121
55
  - Datasets 2.20.0
56
  - Tokenizers 0.19.1
 
1
  ---
2
  license: mit
3
+ base_model: facebook/mbart-large-50
4
  tags:
 
5
  - generated_from_trainer
6
+ metrics:
7
+ - bleu
8
  model-index:
9
+ - name: mbart-simplificacion
10
  results: []
11
+ datasets:
12
+ - crrodrvi/lectura_dificil_facil
13
+ language:
14
+ - es
15
+ pipeline_tag: text2text-generation
16
  ---
 
 
 
 
17
  # modelo_simplificacion_bart
18
 
19
+ This model is a fine-tuned version of [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) on the lectura_dificl_facil dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 2.9828
22
+ - Bleu: 9.1817
23
 
24
  ## Model description
25
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 5.6e-05
42
  - train_batch_size: 4
43
  - eval_batch_size: 4
44
  - seed: 42
 
 
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
+ - num_epochs: 3
 
48
 
49
  ### Training results
50
 
51
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
52
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:-------:|
53
+ | No log | 1.0 | 109 | 3.1239 | 8.7241 | 29.2727 |
54
+ | No log | 2.0 | 218 | 3.0680 | 7.3130 | 23.9432 |
55
+ | No log | 3.0 | 327 | 2.9828 | 9.1817 | 28.5000 |
56
+
57
 
58
 
59
  ### Framework versions
60
 
61
+ - Transformers 4.41.2
62
+ - Pytorch 2.3.0+cu121
63
  - Datasets 2.20.0
64
  - Tokenizers 0.19.1