hamzawaheed commited on
Commit
44599a6
·
verified ·
1 Parent(s): 53d6af2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +3 -12
README.md CHANGED
@@ -23,12 +23,12 @@ The **Emotion Classification Model** is a fine-tuned version of the `distilbert-
23
 
24
  This model leverages the pre-trained language understanding capabilities of DistilBERT to accurately categorize textual data into the following emotion classes:
25
 
26
- - **Joy**
27
  - **Sadness**
 
 
28
  - **Anger**
29
  - **Fear**
30
  - **Surprise**
31
- - **Disgust**
32
 
33
  By fine-tuning on the `dair-ai/emotion` dataset, the model has been optimized to recognize and differentiate subtle emotional cues in various text inputs, making it suitable for applications that require nuanced sentiment analysis and emotional intelligence.
34
 
@@ -39,7 +39,6 @@ By fine-tuning on the `dair-ai/emotion` dataset, the model has been optimized to
39
  The Emotion Classification Model is designed for a variety of applications where understanding the emotional tone of text is crucial. Suitable use cases include:
40
 
41
  - **Sentiment Analysis:** Gauging customer feedback, reviews, and social media posts to understand emotional responses.
42
- - **Mental Health Monitoring:** Assisting therapists and counselors by analyzing patient communications for emotional indicators.
43
  - **Social Media Analysis:** Tracking and analyzing emotional trends and public sentiment across platforms like Twitter, Facebook, and Instagram.
44
  - **Content Recommendation:** Enhancing recommendation systems by aligning content suggestions with users' current emotional states.
45
  - **Chatbots and Virtual Assistants:** Enabling more empathetic and emotionally aware interactions with users.
@@ -51,7 +50,7 @@ While the Emotion Classification Model demonstrates strong performance across va
51
  - **Bias in Training Data:** The model may inherit biases present in the `dair-ai/emotion` dataset, potentially affecting its performance across different demographics, cultures, or contexts.
52
  - **Contextual Understanding:** The model analyzes text in isolation and may struggle with understanding nuanced emotions that depend on broader conversational context or preceding interactions.
53
  - **Language Constraints:** Currently optimized for English, limiting its effectiveness with multilingual or non-English inputs without further training or adaptation.
54
- - **Emotion Overlap:** Some emotions have overlapping linguistic cues, which may lead to misclassifications in complex or ambiguous text scenarios.
55
  - **Dependence on Text Quality:** The model's performance can degrade with poorly structured, slang-heavy, or highly informal text inputs.
56
 
57
  ## Training and Evaluation Data
@@ -66,13 +65,6 @@ The model was trained and evaluated on the [`dair-ai/emotion`](https://huggingfa
66
  - **Training Set:** 16,000 samples
67
  - **Validation Set:** 2,000 samples
68
  - **Test Set:** 2,000 samples
69
- - **Emotion Classes:** 6
70
- - **Joy:** 3,000 samples
71
- - **Sadness:** 3,500 samples
72
- - **Anger:** 2,500 samples
73
- - **Fear:** 2,000 samples
74
- - **Surprise:** 4,000 samples
75
- - **Disgust:** 2,000 samples
76
 
77
  ### Data Preprocessing
78
 
@@ -107,7 +99,6 @@ The following hyperparameters were used during training:
107
 
108
  - **Mixed Precision Training:** Utilized PyTorch's Native AMP to accelerate training and reduce memory consumption when a CUDA-enabled GPU is available.
109
  - **Gradient Accumulation:** Implemented gradient accumulation with `2` steps to effectively increase the batch size without exceeding GPU memory limits.
110
- - **Early Stopping:** Incorporated `EarlyStoppingCallback` with a patience of `2` epochs to halt training if the validation loss does not improve, preventing overfitting.
111
  - **Checkpointing:** Configured to save model checkpoints at the end of each epoch, retaining only the two most recent checkpoints to manage storage efficiently.
112
 
113
  ### Training Duration
 
23
 
24
  This model leverages the pre-trained language understanding capabilities of DistilBERT to accurately categorize textual data into the following emotion classes:
25
 
 
26
  - **Sadness**
27
+ - **Joy**
28
+ - **Love**
29
  - **Anger**
30
  - **Fear**
31
  - **Surprise**
 
32
 
33
  By fine-tuning on the `dair-ai/emotion` dataset, the model has been optimized to recognize and differentiate subtle emotional cues in various text inputs, making it suitable for applications that require nuanced sentiment analysis and emotional intelligence.
34
 
 
39
  The Emotion Classification Model is designed for a variety of applications where understanding the emotional tone of text is crucial. Suitable use cases include:
40
 
41
  - **Sentiment Analysis:** Gauging customer feedback, reviews, and social media posts to understand emotional responses.
 
42
  - **Social Media Analysis:** Tracking and analyzing emotional trends and public sentiment across platforms like Twitter, Facebook, and Instagram.
43
  - **Content Recommendation:** Enhancing recommendation systems by aligning content suggestions with users' current emotional states.
44
  - **Chatbots and Virtual Assistants:** Enabling more empathetic and emotionally aware interactions with users.
 
50
  - **Bias in Training Data:** The model may inherit biases present in the `dair-ai/emotion` dataset, potentially affecting its performance across different demographics, cultures, or contexts.
51
  - **Contextual Understanding:** The model analyzes text in isolation and may struggle with understanding nuanced emotions that depend on broader conversational context or preceding interactions.
52
  - **Language Constraints:** Currently optimized for English, limiting its effectiveness with multilingual or non-English inputs without further training or adaptation.
53
+ - **Emotion Overlap:** Some emotions have overlapping linguistic cues, which may lead to misclassifications in ambiguous text scenarios.
54
  - **Dependence on Text Quality:** The model's performance can degrade with poorly structured, slang-heavy, or highly informal text inputs.
55
 
56
  ## Training and Evaluation Data
 
65
  - **Training Set:** 16,000 samples
66
  - **Validation Set:** 2,000 samples
67
  - **Test Set:** 2,000 samples
 
 
 
 
 
 
 
68
 
69
  ### Data Preprocessing
70
 
 
99
 
100
  - **Mixed Precision Training:** Utilized PyTorch's Native AMP to accelerate training and reduce memory consumption when a CUDA-enabled GPU is available.
101
  - **Gradient Accumulation:** Implemented gradient accumulation with `2` steps to effectively increase the batch size without exceeding GPU memory limits.
 
102
  - **Checkpointing:** Configured to save model checkpoints at the end of each epoch, retaining only the two most recent checkpoints to manage storage efficiently.
103
 
104
  ### Training Duration