End of training
Browse files- README.md +69 -0
- config.json +40 -0
- merges.txt +0 -0
- model.safetensors +3 -0
- runs/Jan24_14-37-05_579ed4a51ca4/events.out.tfevents.1706107026.579ed4a51ca4.457.0 +3 -0
- runs/Jan24_14-37-05_579ed4a51ca4/events.out.tfevents.1706107294.579ed4a51ca4.457.1 +3 -0
- runs/Jan24_17-57-00_7d2b1e2d4e1a/events.out.tfevents.1706119022.7d2b1e2d4e1a.2199.0 +3 -0
- runs/Jan24_17-57-30_7d2b1e2d4e1a/events.out.tfevents.1706119051.7d2b1e2d4e1a.2199.1 +3 -0
- runs/Jan24_17-57-30_7d2b1e2d4e1a/events.out.tfevents.1706119493.7d2b1e2d4e1a.2199.3 +3 -0
- runs/Jan24_17-59-31_7d2b1e2d4e1a/events.out.tfevents.1706119172.7d2b1e2d4e1a.2199.2 +3 -0
- runs/Jan24_17-59-31_7d2b1e2d4e1a/events.out.tfevents.1706119503.7d2b1e2d4e1a.2199.4 +3 -0
- runs/Jan24_18-39-04_7d2b1e2d4e1a/events.out.tfevents.1706121545.7d2b1e2d4e1a.2199.7 +3 -0
- runs/Jan24_19-20-11_7d2b1e2d4e1a/events.out.tfevents.1706124012.7d2b1e2d4e1a.2199.8 +3 -0
- runs/Jan24_19-23-12_7d2b1e2d4e1a/events.out.tfevents.1706124192.7d2b1e2d4e1a.2199.9 +3 -0
- runs/Jan24_19-23-37_7d2b1e2d4e1a/events.out.tfevents.1706124217.7d2b1e2d4e1a.2199.10 +3 -0
- runs/Jan24_19-31-14_7d2b1e2d4e1a/events.out.tfevents.1706124675.7d2b1e2d4e1a.26662.0 +3 -0
- runs/Jan24_19-32-18_7d2b1e2d4e1a/events.out.tfevents.1706124738.7d2b1e2d4e1a.26662.1 +3 -0
- runs/Jan24_19-33-23_7d2b1e2d4e1a/events.out.tfevents.1706124804.7d2b1e2d4e1a.26662.2 +3 -0
- runs/Jan24_19-33-53_7d2b1e2d4e1a/events.out.tfevents.1706124834.7d2b1e2d4e1a.26662.3 +3 -0
- runs/Jan24_19-39-23_7d2b1e2d4e1a/events.out.tfevents.1706125164.7d2b1e2d4e1a.26662.4 +3 -0
- special_tokens_map.json +6 -0
- tokenizer.json +0 -0
- tokenizer_config.json +20 -0
- training_args.bin +3 -0
- vocab.json +0 -0
- vocab.txt +0 -0
README.md
ADDED
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
base_model: gpt2
|
4 |
+
tags:
|
5 |
+
- generated_from_trainer
|
6 |
+
metrics:
|
7 |
+
- accuracy
|
8 |
+
- f1
|
9 |
+
model-index:
|
10 |
+
- name: SentimentT2_GPT2
|
11 |
+
results: []
|
12 |
+
---
|
13 |
+
|
14 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
15 |
+
should probably proofread and complete it, then remove this comment. -->
|
16 |
+
|
17 |
+
# SentimentT2_GPT2
|
18 |
+
|
19 |
+
This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
|
20 |
+
It achieves the following results on the evaluation set:
|
21 |
+
- Loss: 0.7249
|
22 |
+
- Accuracy: 0.8669
|
23 |
+
- F1: 0.8749
|
24 |
+
- Auc Roc: 0.9317
|
25 |
+
- Log Loss: 0.7249
|
26 |
+
|
27 |
+
## Model description
|
28 |
+
|
29 |
+
More information needed
|
30 |
+
|
31 |
+
## Intended uses & limitations
|
32 |
+
|
33 |
+
More information needed
|
34 |
+
|
35 |
+
## Training and evaluation data
|
36 |
+
|
37 |
+
More information needed
|
38 |
+
|
39 |
+
## Training procedure
|
40 |
+
|
41 |
+
### Training hyperparameters
|
42 |
+
|
43 |
+
The following hyperparameters were used during training:
|
44 |
+
- learning_rate: 1e-05
|
45 |
+
- train_batch_size: 1
|
46 |
+
- eval_batch_size: 1
|
47 |
+
- seed: 42
|
48 |
+
- gradient_accumulation_steps: 2
|
49 |
+
- total_train_batch_size: 2
|
50 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
51 |
+
- lr_scheduler_type: cosine
|
52 |
+
- lr_scheduler_warmup_steps: 500
|
53 |
+
- num_epochs: 2
|
54 |
+
- mixed_precision_training: Native AMP
|
55 |
+
|
56 |
+
### Training results
|
57 |
+
|
58 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Auc Roc | Log Loss |
|
59 |
+
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:-------:|:--------:|
|
60 |
+
| 1.1736 | 1.0 | 1618 | 0.6146 | 0.8507 | 0.8592 | 0.9256 | 0.6146 |
|
61 |
+
| 0.6452 | 2.0 | 3236 | 0.7249 | 0.8669 | 0.8749 | 0.9317 | 0.7249 |
|
62 |
+
|
63 |
+
|
64 |
+
### Framework versions
|
65 |
+
|
66 |
+
- Transformers 4.35.2
|
67 |
+
- Pytorch 2.1.0+cu121
|
68 |
+
- Datasets 2.16.1
|
69 |
+
- Tokenizers 0.15.1
|
config.json
ADDED
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "gpt2",
|
3 |
+
"activation_function": "gelu_new",
|
4 |
+
"architectures": [
|
5 |
+
"GPT2ForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attn_pdrop": 0.1,
|
8 |
+
"bos_token_id": 50256,
|
9 |
+
"embd_pdrop": 0.1,
|
10 |
+
"eos_token_id": 50256,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"layer_norm_epsilon": 1e-05,
|
13 |
+
"model_type": "gpt2",
|
14 |
+
"n_ctx": 1024,
|
15 |
+
"n_embd": 768,
|
16 |
+
"n_head": 12,
|
17 |
+
"n_inner": null,
|
18 |
+
"n_layer": 12,
|
19 |
+
"n_positions": 1024,
|
20 |
+
"problem_type": "single_label_classification",
|
21 |
+
"reorder_and_upcast_attn": false,
|
22 |
+
"resid_pdrop": 0.1,
|
23 |
+
"scale_attn_by_inverse_layer_idx": false,
|
24 |
+
"scale_attn_weights": true,
|
25 |
+
"summary_activation": null,
|
26 |
+
"summary_first_dropout": 0.1,
|
27 |
+
"summary_proj_to_labels": true,
|
28 |
+
"summary_type": "cls_index",
|
29 |
+
"summary_use_proj": true,
|
30 |
+
"task_specific_params": {
|
31 |
+
"text-generation": {
|
32 |
+
"do_sample": true,
|
33 |
+
"max_length": 50
|
34 |
+
}
|
35 |
+
},
|
36 |
+
"torch_dtype": "float32",
|
37 |
+
"transformers_version": "4.35.2",
|
38 |
+
"use_cache": true,
|
39 |
+
"vocab_size": 50257
|
40 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02f0b5d0a85a3b873b4a183a19c22b0acf256a4c08e0f9b1355b1c34211241e5
|
3 |
+
size 497780432
|
runs/Jan24_14-37-05_579ed4a51ca4/events.out.tfevents.1706107026.579ed4a51ca4.457.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9793fe0c6556e907b1c614be0ace204cb95291a8b77b8a77fb2966750c729a64
|
3 |
+
size 7042
|
runs/Jan24_14-37-05_579ed4a51ca4/events.out.tfevents.1706107294.579ed4a51ca4.457.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a839574f814954dbb18f4b789b2bf531497e649d8830ae90a2484a723adc8d1e
|
3 |
+
size 560
|
runs/Jan24_17-57-00_7d2b1e2d4e1a/events.out.tfevents.1706119022.7d2b1e2d4e1a.2199.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cd2d584b67487707bc9efea3ec34eb13466a73d8f15f61b3954cc69b9c0715d8
|
3 |
+
size 4184
|
runs/Jan24_17-57-30_7d2b1e2d4e1a/events.out.tfevents.1706119051.7d2b1e2d4e1a.2199.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ade5df7e518384ed3538208ee6c86b7c11214a6eceaa75b0e2a65effdbec6729
|
3 |
+
size 5784
|
runs/Jan24_17-57-30_7d2b1e2d4e1a/events.out.tfevents.1706119493.7d2b1e2d4e1a.2199.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5b3214ab547043011326ad1369c0d723f88faf092bab310c987101761c07b531
|
3 |
+
size 560
|
runs/Jan24_17-59-31_7d2b1e2d4e1a/events.out.tfevents.1706119172.7d2b1e2d4e1a.2199.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1a61710090d075321d12b7f8cacff814f10e9b8b64a06b7373eadc10dd1dee00
|
3 |
+
size 5905
|
runs/Jan24_17-59-31_7d2b1e2d4e1a/events.out.tfevents.1706119503.7d2b1e2d4e1a.2199.4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ebeb647374b9b6ed5833b093f646e9c089c5a8ac9c0be5fd010d01d22832d1ca
|
3 |
+
size 560
|
runs/Jan24_18-39-04_7d2b1e2d4e1a/events.out.tfevents.1706121545.7d2b1e2d4e1a.2199.7
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2aa18069c9bab305333570793f1aaab2ebe626ed027930529345da1d81bd1924
|
3 |
+
size 10938
|
runs/Jan24_19-20-11_7d2b1e2d4e1a/events.out.tfevents.1706124012.7d2b1e2d4e1a.2199.8
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e1ff5cea8170bbdd6e49c9181719bfa37e826252a7912b96c5bfd3379d9ac8e9
|
3 |
+
size 4544
|
runs/Jan24_19-23-12_7d2b1e2d4e1a/events.out.tfevents.1706124192.7d2b1e2d4e1a.2199.9
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d48965daa51cd8c292e9726fa9ce868c27b7b7a0b26953610a1284b0a3244db0
|
3 |
+
size 4544
|
runs/Jan24_19-23-37_7d2b1e2d4e1a/events.out.tfevents.1706124217.7d2b1e2d4e1a.2199.10
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92ce18dee6663fdd82856765199eafdf48f24704820cb05dea5b7f9a1d3e3b4d
|
3 |
+
size 4543
|
runs/Jan24_19-31-14_7d2b1e2d4e1a/events.out.tfevents.1706124675.7d2b1e2d4e1a.26662.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f052f9d74b7d2a9b38e8f79422b89a997f4df9ad34e859927deefd6aeb99e59c
|
3 |
+
size 4543
|
runs/Jan24_19-32-18_7d2b1e2d4e1a/events.out.tfevents.1706124738.7d2b1e2d4e1a.26662.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:983fea0d1e6ed2f5a2d7d497e5a5e25f161ea62d41b9150f99dd72d0af6a2211
|
3 |
+
size 4543
|
runs/Jan24_19-33-23_7d2b1e2d4e1a/events.out.tfevents.1706124804.7d2b1e2d4e1a.26662.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d826890967001774b0ee65bd6c10a7a5ac575902f6bff7de0c7a97a409809e58
|
3 |
+
size 4543
|
runs/Jan24_19-33-53_7d2b1e2d4e1a/events.out.tfevents.1706124834.7d2b1e2d4e1a.26662.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53026ad3317456be6c8ffc66d5cc581d9e09640d1bfb7ae3af5c1f947ca13845
|
3 |
+
size 4700
|
runs/Jan24_19-39-23_7d2b1e2d4e1a/events.out.tfevents.1706125164.7d2b1e2d4e1a.26662.4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7f0106f7873903e963089ccc4d13d050b3100e0bccfd2c4fb500c1e0835913c4
|
3 |
+
size 6154
|
special_tokens_map.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<|endoftext|>",
|
3 |
+
"eos_token": "<|endoftext|>",
|
4 |
+
"pad_token": "<|endoftext|>",
|
5 |
+
"unk_token": "<|endoftext|>"
|
6 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"50256": {
|
5 |
+
"content": "<|endoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
}
|
12 |
+
},
|
13 |
+
"bos_token": "<|endoftext|>",
|
14 |
+
"clean_up_tokenization_spaces": true,
|
15 |
+
"eos_token": "<|endoftext|>",
|
16 |
+
"model_max_length": 1024,
|
17 |
+
"pad_token": "<|endoftext|>",
|
18 |
+
"tokenizer_class": "GPT2Tokenizer",
|
19 |
+
"unk_token": "<|endoftext|>"
|
20 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3de59719afd225fb332ed6d1b42b6d228a99601e6a3b5db147d421b6288ddab6
|
3 |
+
size 4664
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|