Commit
·
75f6c1d
1
Parent(s):
6dfb2d8
Update README.md
Browse files
README.md
CHANGED
@@ -18,7 +18,19 @@ should probably proofread and complete it, then remove this comment. -->
|
|
18 |
# t5_recommendation_sports_equipment_english
|
19 |
|
20 |
This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on a custom dataset, consisting of sports equipment customers have purchased, and items to recommended next.
|
21 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
- Loss: 0.4554
|
23 |
- Rouge1: 57.1429
|
24 |
- Rouge2: 47.6190
|
@@ -28,18 +40,24 @@ It achieves the following results on the evaluation set:
|
|
28 |
|
29 |
## Model description
|
30 |
|
31 |
-
|
|
|
|
|
|
|
|
|
32 |
|
33 |
## Intended uses & limitations
|
34 |
|
35 |
-
|
36 |
|
37 |
## Training and evaluation data
|
38 |
|
39 |
-
|
40 |
|
41 |
## Training procedure
|
42 |
|
|
|
|
|
43 |
### Training hyperparameters
|
44 |
|
45 |
The following hyperparameters were used during training:
|
|
|
18 |
# t5_recommendation_sports_equipment_english
|
19 |
|
20 |
This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on a custom dataset, consisting of sports equipment customers have purchased, and items to recommended next.
|
21 |
+
|
22 |
+
This is based on the paper ["Recommendation as Language Processing (RLP): A Unified Pretrain, Personalized Prompt & Predict Paradigm (P5)"](https://arxiv.org/pdf/2203.13366.pdf), where the researchers use a language model as a recommendation system.
|
23 |
+
- LLMs can "understand" relationships between words/terms via embeddings produced by the transformer architecture. This allows for relationships to be taken into account.
|
24 |
+
- By feeding an LLM a history of items purchased as the input and the next item purchased as the output, the model can learn what to recommend based on the semantics of the product's name.
|
25 |
+
- Taking multiple examples of different users' purchase history into account, the LLM can also learn what genres of products go with what.
|
26 |
+
- This essentially replicates collaboritve filtering
|
27 |
+
- Benefits include:
|
28 |
+
- Getting past the cold-start problem with ease (when new items are introduced, the model will be able to understand what's similar to it from the name alone).
|
29 |
+
- Avoiding tedious, manual feature engineering (using natural language, the LLM will automatically learn).
|
30 |
+
|
31 |
+
The github repository for fine-tuning this model can be viewed [here](https://github.com/Mohammadhia/t5_p5_recommendation_system).
|
32 |
+
|
33 |
+
The fine-tuned T5 model achieves the following results on the evaluation set:
|
34 |
- Loss: 0.4554
|
35 |
- Rouge1: 57.1429
|
36 |
- Rouge2: 47.6190
|
|
|
40 |
|
41 |
## Model description
|
42 |
|
43 |
+
T5 is an open-source sequence-to-sequence model released by Google in 2020, from which several variants have been developed. This fine-tuned version is an attempt to replicate what was presented in the [P5 paper](https://arxiv.org/pdf/2203.13366.pdf), with a custom dataset (based on sports equipment).
|
44 |
+
|
45 |
+
More about this model (T5) can be viewed [here](https://huggingface.co/docs/transformers/model_doc/t5).
|
46 |
+
|
47 |
+
The P5 models from the paper can be viewed on the [Hugging Face Hub](https://huggingface.co/makitanikaze/P5) as well as in this [repository](https://github.com/jeykigung/P5).
|
48 |
|
49 |
## Intended uses & limitations
|
50 |
|
51 |
+
Can be used as you please, but is limited to the sports equipment dataset it was fine-tuned on. Your mileage may vary.
|
52 |
|
53 |
## Training and evaluation data
|
54 |
|
55 |
+
Please see this [repository](https://github.com/Mohammadhia/t5_p5_recommendation_system) for training and evaluation data.
|
56 |
|
57 |
## Training procedure
|
58 |
|
59 |
+
Please see this [repository](https://github.com/Mohammadhia/t5_p5_recommendation_system) for training and evaluation data.
|
60 |
+
|
61 |
### Training hyperparameters
|
62 |
|
63 |
The following hyperparameters were used during training:
|