Advince commited on
Commit
fa48b02
·
verified ·
1 Parent(s): 0c82345

Advince/distilbert-base-uncased-lora-toxic-classification

Browse files
README.md CHANGED
@@ -1,15 +1,15 @@
1
- ---
2
- base_model: distilbert-base-uncased
3
- library_name: peft
4
- license: apache-2.0
5
- metrics:
6
- - accuracy
7
- tags:
8
- - generated_from_trainer
9
- model-index:
10
- - name: distilbert-base-uncased-lora-text-classification
11
- results: []
12
- ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 1.3115
22
- - Accuracy: {'accuracy': 0.8125}
23
 
24
  ## Model description
25
 
@@ -50,27 +50,27 @@ The following hyperparameters were used during training:
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------------------------------:|
53
- | No log | 1.0 | 431 | 0.6279 | {'accuracy': 0.7135416666666666} |
54
- | 0.7248 | 2.0 | 862 | 0.5949 | {'accuracy': 0.7604166666666666} |
55
- | 0.5433 | 3.0 | 1293 | 0.5986 | {'accuracy': 0.8072916666666666} |
56
- | 0.4762 | 4.0 | 1724 | 0.6967 | {'accuracy': 0.8125} |
57
- | 0.3971 | 5.0 | 2155 | 0.7136 | {'accuracy': 0.8229166666666666} |
58
- | 0.3471 | 6.0 | 2586 | 0.8597 | {'accuracy': 0.8177083333333334} |
59
- | 0.2695 | 7.0 | 3017 | 1.0061 | {'accuracy': 0.8072916666666666} |
60
- | 0.2695 | 8.0 | 3448 | 0.7674 | {'accuracy': 0.8333333333333334} |
61
- | 0.2417 | 9.0 | 3879 | 1.2479 | {'accuracy': 0.828125} |
62
- | 0.2079 | 10.0 | 4310 | 1.0548 | {'accuracy': 0.8177083333333334} |
63
- | 0.1941 | 11.0 | 4741 | 1.0516 | {'accuracy': 0.8229166666666666} |
64
- | 0.1711 | 12.0 | 5172 | 1.2246 | {'accuracy': 0.828125} |
65
- | 0.1253 | 13.0 | 5603 | 1.2416 | {'accuracy': 0.8177083333333334} |
66
- | 0.0918 | 14.0 | 6034 | 1.3199 | {'accuracy': 0.8229166666666666} |
67
- | 0.0918 | 15.0 | 6465 | 1.3115 | {'accuracy': 0.8125} |
68
 
69
 
70
  ### Framework versions
71
 
72
- - PEFT 0.13.0
73
  - Transformers 4.44.2
74
- - Pytorch 2.4.1+cu121
75
  - Datasets 3.0.0
76
  - Tokenizers 0.19.1
 
1
+ ---
2
+ base_model: distilbert-base-uncased
3
+ library_name: peft
4
+ license: apache-2.0
5
+ metrics:
6
+ - accuracy
7
+ tags:
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: distilbert-base-uncased-lora-text-classification
11
+ results: []
12
+ ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
 
18
 
19
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.5043
22
+ - Accuracy: {'accuracy': 0.9479166666666666}
23
 
24
  ## Model description
25
 
 
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------------------------------:|
53
+ | No log | 1.0 | 431 | 0.2331 | {'accuracy': 0.90625} |
54
+ | 0.4451 | 2.0 | 862 | 0.3140 | {'accuracy': 0.90625} |
55
+ | 0.2963 | 3.0 | 1293 | 0.3216 | {'accuracy': 0.9322916666666666} |
56
+ | 0.25 | 4.0 | 1724 | 0.2690 | {'accuracy': 0.9270833333333334} |
57
+ | 0.2261 | 5.0 | 2155 | 0.2707 | {'accuracy': 0.9479166666666666} |
58
+ | 0.1511 | 6.0 | 2586 | 0.2543 | {'accuracy': 0.9427083333333334} |
59
+ | 0.1401 | 7.0 | 3017 | 0.3120 | {'accuracy': 0.9375} |
60
+ | 0.1401 | 8.0 | 3448 | 0.2845 | {'accuracy': 0.953125} |
61
+ | 0.086 | 9.0 | 3879 | 0.4018 | {'accuracy': 0.921875} |
62
+ | 0.0583 | 10.0 | 4310 | 0.4593 | {'accuracy': 0.9427083333333334} |
63
+ | 0.0475 | 11.0 | 4741 | 0.4401 | {'accuracy': 0.953125} |
64
+ | 0.0515 | 12.0 | 5172 | 0.4631 | {'accuracy': 0.9479166666666666} |
65
+ | 0.0291 | 13.0 | 5603 | 0.4593 | {'accuracy': 0.9479166666666666} |
66
+ | 0.0319 | 14.0 | 6034 | 0.5292 | {'accuracy': 0.9479166666666666} |
67
+ | 0.0319 | 15.0 | 6465 | 0.5043 | {'accuracy': 0.9479166666666666} |
68
 
69
 
70
  ### Framework versions
71
 
72
+ - PEFT 0.12.0
73
  - Transformers 4.44.2
74
+ - Pytorch 2.4.1+cu124
75
  - Datasets 3.0.0
76
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -1,31 +1,31 @@
1
- {
2
- "alpha_pattern": {},
3
- "auto_mapping": null,
4
- "base_model_name_or_path": "distilbert-base-uncased",
5
- "bias": "none",
6
- "fan_in_fan_out": false,
7
- "inference_mode": true,
8
- "init_lora_weights": true,
9
- "layer_replication": null,
10
- "layers_pattern": null,
11
- "layers_to_transform": null,
12
- "loftq_config": {},
13
- "lora_alpha": 32,
14
- "lora_dropout": 0.01,
15
- "megatron_config": null,
16
- "megatron_core": "megatron.core",
17
- "modules_to_save": [
18
- "classifier",
19
- "score"
20
- ],
21
- "peft_type": "LORA",
22
- "r": 4,
23
- "rank_pattern": {},
24
- "revision": null,
25
- "target_modules": [
26
- "q_lin"
27
- ],
28
- "task_type": "SEQ_CLS",
29
- "use_dora": false,
30
- "use_rslora": false
31
  }
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "distilbert-base-uncased",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.01,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": [
18
+ "classifier",
19
+ "score"
20
+ ],
21
+ "peft_type": "LORA",
22
+ "r": 4,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "q_lin"
27
+ ],
28
+ "task_type": "SEQ_CLS",
29
+ "use_dora": false,
30
+ "use_rslora": false
31
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70c55ad8ee709d5ca29ddaaa939b5c82162f72af76aa36960605b9316e294432
3
- size 2521180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7889b99657a6e509ad9627cd74ec812561083b7486f8a47d3361f190a4ad6db
3
+ size 2518104
runs/Sep21_15-24-52_LAPTOP-9511T4P2/events.out.tfevents.1726925094.LAPTOP-9511T4P2.31280.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6f7372a406c70e8453a3b676d89538b19cbd0f17ce3bc318c19d2cbe4458ac4
3
+ size 5008
runs/Sep21_15-28-50_LAPTOP-9511T4P2/events.out.tfevents.1726925331.LAPTOP-9511T4P2.5348.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5477a3950e83a6c1f34ab56caa4cd048e460ab15cc371105bbdd329652dcd08a
3
+ size 6032
runs/Sep23_16-42-58_LAPTOP-9511T4P2/events.out.tfevents.1727102580.LAPTOP-9511T4P2.15204.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46228cee720331ca19332d3484b96622e60e83211a82800e396c0b47246e53c5
3
+ size 5004
special_tokens_map.json CHANGED
@@ -1,7 +1,7 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer_config.json CHANGED
@@ -1,56 +1,56 @@
1
- {
2
- "add_prefix_space": true,
3
- "added_tokens_decoder": {
4
- "0": {
5
- "content": "[PAD]",
6
- "lstrip": false,
7
- "normalized": false,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": true
11
- },
12
- "100": {
13
- "content": "[UNK]",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- },
20
- "101": {
21
- "content": "[CLS]",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false,
26
- "special": true
27
- },
28
- "102": {
29
- "content": "[SEP]",
30
- "lstrip": false,
31
- "normalized": false,
32
- "rstrip": false,
33
- "single_word": false,
34
- "special": true
35
- },
36
- "103": {
37
- "content": "[MASK]",
38
- "lstrip": false,
39
- "normalized": false,
40
- "rstrip": false,
41
- "single_word": false,
42
- "special": true
43
- }
44
- },
45
- "clean_up_tokenization_spaces": true,
46
- "cls_token": "[CLS]",
47
- "do_lower_case": true,
48
- "mask_token": "[MASK]",
49
- "model_max_length": 512,
50
- "pad_token": "[PAD]",
51
- "sep_token": "[SEP]",
52
- "strip_accents": null,
53
- "tokenize_chinese_chars": true,
54
- "tokenizer_class": "DistilBertTokenizer",
55
- "unk_token": "[UNK]"
56
- }
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "[PAD]",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "100": {
13
+ "content": "[UNK]",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "101": {
21
+ "content": "[CLS]",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "102": {
29
+ "content": "[SEP]",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "103": {
37
+ "content": "[MASK]",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "DistilBertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66153193613f4463045ef8aba5885b66b724bc83111c9867a7656c6db070c4a2
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bffcb5a4cd3f18e69158eb5d87cf3cb5008efd9551c2f305ebe8600e39e3fad1
3
  size 5240