Update README.md
Browse files
README.md
CHANGED
@@ -2,17 +2,80 @@
|
|
2 |
license: apache-2.0
|
3 |
tags:
|
4 |
- generated_from_trainer
|
5 |
-
|
6 |
-
-
|
7 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
---
|
9 |
|
10 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
11 |
-
should probably proofread and complete it, then remove this comment. -->
|
12 |
|
13 |
-
# distilgpt2-emailgen
|
14 |
|
15 |
-
This
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
It achieves the following results on the evaluation set:
|
17 |
- Loss: 1.9126
|
18 |
|
@@ -30,7 +93,11 @@ More information needed
|
|
30 |
|
31 |
## Training procedure
|
32 |
|
33 |
-
### Training hyperparameters
|
|
|
|
|
|
|
|
|
34 |
|
35 |
The following hyperparameters were used during training:
|
36 |
- learning_rate: 0.0006
|
|
|
2 |
license: apache-2.0
|
3 |
tags:
|
4 |
- generated_from_trainer
|
5 |
+
- distilgpt2
|
6 |
+
- email generation
|
7 |
+
- email
|
8 |
+
datasets:
|
9 |
+
- aeslc
|
10 |
+
- postbot/multi-emails-100k
|
11 |
+
|
12 |
+
widget:
|
13 |
+
- text: "Good Morning Professor Beans,
|
14 |
+
|
15 |
+
Hope you are doing well. I just wanted to reach out and ask if differential calculus will be on the exam"
|
16 |
+
example_title: "email to prof"
|
17 |
+
- text: "Hey <NAME>,\n\nThank you for signing up for my weekly newsletter. Before we get started, you'll have to confirm your email address."
|
18 |
+
example_title: "newsletter"
|
19 |
+
- text: "Hi <NAME>,\n\nI hope this email finds you well. I wanted to reach out and ask about office hours"
|
20 |
+
example_title: "office hours"
|
21 |
+
- text: "Greetings <NAME>,\n\nI hope you had a splendid evening at the Company sausage eating festival. I am reaching out because"
|
22 |
+
example_title: "festival"
|
23 |
+
- text: "Good Morning Harold,\n\nI was wondering when the next"
|
24 |
+
example_title: "event"
|
25 |
+
- text: "URGENT - I need the TPS reports"
|
26 |
+
example_title: "URGENT"
|
27 |
+
- text: "Hi Archibald,\n\nI hope this email finds you extremely well."
|
28 |
+
example_title: "emails that find you"
|
29 |
+
- text: "Hello there.\n\nI just wanted to reach out and check in to"
|
30 |
+
example_title: "checking in"
|
31 |
+
- text: "Hello <NAME>,\n\nI hope this email finds you well. I wanted to reach out and see if you've enjoyed your time with us"
|
32 |
+
example_title: "work well"
|
33 |
+
- text: "Hi <NAME>,\n\nI hope this email finds you well. I wanted to reach out and see if we could catch up"
|
34 |
+
example_title: "catch up"
|
35 |
+
- text: "I'm <NAME> and I just moved into the area and wanted to reach out and get some details on where I could get groceries and"
|
36 |
+
example_title: "grocery"
|
37 |
+
parameters:
|
38 |
+
min_length: 4
|
39 |
+
max_length: 128
|
40 |
+
length_penalty: 0.8
|
41 |
+
no_repeat_ngram_size: 2
|
42 |
+
do_sample: False
|
43 |
+
num_beams: 8
|
44 |
+
early_stopping: True
|
45 |
+
repetition_penalty: 5.5
|
46 |
---
|
47 |
|
|
|
|
|
48 |
|
49 |
+
# distilgpt2-emailgen: V2
|
50 |
|
51 |
+
> This is a V2, which should perform better than V1. This is in the process of being evaluated.
|
52 |
+
|
53 |
+
Why write the rest of your email when you can generate it?
|
54 |
+
|
55 |
+
```python
|
56 |
+
from transformers import pipeline
|
57 |
+
|
58 |
+
model_tag = "postbot/distilgpt2-emailgen"
|
59 |
+
generator = pipeline(
|
60 |
+
'text-generation',
|
61 |
+
model=model_tag,
|
62 |
+
)
|
63 |
+
|
64 |
+
prompt = """
|
65 |
+
Hello,
|
66 |
+
|
67 |
+
Following up on the bubblegum shipment."""
|
68 |
+
|
69 |
+
result = generator(
|
70 |
+
prompt,
|
71 |
+
max_length=64,
|
72 |
+
do_sample=False,
|
73 |
+
early_stopping=True,
|
74 |
+
) # generate
|
75 |
+
print(result[0]['generated_text'])
|
76 |
+
```
|
77 |
+
|
78 |
+
This model is a fine-tuned version of [postbot/distilgpt2-emailgen-V2](https://huggingface.co/postbot/distilgpt2-emailgen-V2) on the postbot/multi-emails-100k dataset.
|
79 |
It achieves the following results on the evaluation set:
|
80 |
- Loss: 1.9126
|
81 |
|
|
|
93 |
|
94 |
## Training procedure
|
95 |
|
96 |
+
### Training hyperparameters (run 1/2)
|
97 |
+
|
98 |
+
TODO
|
99 |
+
|
100 |
+
### Training hyperparameters (run 2/2)
|
101 |
|
102 |
The following hyperparameters were used during training:
|
103 |
- learning_rate: 0.0006
|