vincentmin commited on
Commit
2a74db6
·
1 Parent(s): 9f2aaea

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -20
README.md CHANGED
@@ -2,10 +2,11 @@
2
  base_model: meta-llama/Llama-2-7b-chat-hf
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: llama-2-7b-reward-oasst1
7
  results: []
8
- library_name: peft
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,13 +16,8 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 0.5659
19
- - eval_accuracy: 0.7458
20
- - eval_runtime: 435.7582
21
- - eval_samples_per_second: 2.04
22
- - eval_steps_per_second: 2.04
23
- - epoch: 0.7
24
- - step: 1750
25
 
26
  ## Model description
27
 
@@ -37,17 +33,6 @@ More information needed
37
 
38
  ## Training procedure
39
 
40
-
41
- The following `bitsandbytes` quantization config was used during training:
42
- - load_in_8bit: False
43
- - load_in_4bit: True
44
- - llm_int8_threshold: 6.0
45
- - llm_int8_skip_modules: None
46
- - llm_int8_enable_fp32_cpu_offload: False
47
- - llm_int8_has_fp16_weight: False
48
- - bnb_4bit_quant_type: nf4
49
- - bnb_4bit_use_double_quant: False
50
- - bnb_4bit_compute_dtype: float16
51
  ### Training hyperparameters
52
 
53
  The following hyperparameters were used during training:
@@ -61,9 +46,24 @@ The following hyperparameters were used during training:
61
  - lr_scheduler_type: linear
62
  - num_epochs: 1
63
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
64
  ### Framework versions
65
 
66
- - PEFT 0.5.0.dev0
67
  - Transformers 4.32.0.dev0
68
  - Pytorch 2.0.1+cu118
69
  - Datasets 2.14.0
 
2
  base_model: meta-llama/Llama-2-7b-chat-hf
3
  tags:
4
  - generated_from_trainer
5
+ metrics:
6
+ - accuracy
7
  model-index:
8
  - name: llama-2-7b-reward-oasst1
9
  results: []
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
16
 
17
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.5713
20
+ - Accuracy: 0.7435
 
 
 
 
 
21
 
22
  ## Model description
23
 
 
33
 
34
  ## Training procedure
35
 
 
 
 
 
 
 
 
 
 
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
 
46
  - lr_scheduler_type: linear
47
  - num_epochs: 1
48
 
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
+ | 0.8409 | 0.1 | 250 | 0.8243 | 0.6220 |
54
+ | 0.6288 | 0.2 | 500 | 0.7539 | 0.6715 |
55
+ | 0.5882 | 0.3 | 750 | 0.6792 | 0.7075 |
56
+ | 0.7671 | 0.4 | 1000 | 0.6130 | 0.7334 |
57
+ | 0.5782 | 0.5 | 1250 | 0.6115 | 0.7255 |
58
+ | 0.5691 | 0.6 | 1500 | 0.5795 | 0.7413 |
59
+ | 0.6579 | 0.7 | 1750 | 0.5774 | 0.7469 |
60
+ | 0.6107 | 0.8 | 2000 | 0.5691 | 0.7402 |
61
+ | 0.6255 | 0.9 | 2250 | 0.5710 | 0.7435 |
62
+ | 0.7034 | 1.0 | 2500 | 0.5713 | 0.7435 |
63
+
64
+
65
  ### Framework versions
66
 
 
67
  - Transformers 4.32.0.dev0
68
  - Pytorch 2.0.1+cu118
69
  - Datasets 2.14.0