jennasparks commited on
Commit
ca7705e
·
verified ·
1 Parent(s): fbc8c92

Updated with model info

Browse files
Files changed (1) hide show
  1. README.md +14 -14
README.md CHANGED
@@ -8,23 +8,23 @@ pinned: false
8
  ---
9
 
10
 
11
- # Random Baseline Model for Climate Disinformation Classification
12
 
13
  ## Model Description
14
 
15
- This is a random baseline model for the Frugal AI Challenge 2024, specifically for the text classification task of identifying climate disinformation. The model serves as a performance floor, randomly assigning labels to text inputs without any learning.
16
 
17
  ### Intended Use
18
 
19
- - **Primary intended uses**: Baseline comparison for climate disinformation classification models
20
  - **Primary intended users**: Researchers and developers participating in the Frugal AI Challenge
21
  - **Out-of-scope use cases**: Not intended for production use or real-world classification tasks
22
 
23
  ## Training Data
24
 
25
- The model uses the QuotaClimat/frugalaichallenge-text-train dataset:
26
  - Size: ~6000 examples
27
- - Split: 80% train, 20% test
28
  - 8 categories of climate disinformation claims
29
 
30
  ### Labels
@@ -37,16 +37,19 @@ The model uses the QuotaClimat/frugalaichallenge-text-train dataset:
37
  6. Proponents are biased
38
  7. Fossil fuels are needed
39
 
 
 
 
40
  ## Performance
41
 
42
  ### Metrics
43
- - **Accuracy**: ~12.5% (random chance with 8 classes)
44
  - **Environmental Impact**:
45
  - Emissions tracked in gCO2eq
46
  - Energy consumption tracked in Wh
47
 
48
  ### Model Architecture
49
- The model implements a random choice between the 8 possible labels, serving as the simplest possible baseline.
50
 
51
  ## Environmental Impact
52
 
@@ -54,18 +57,15 @@ Environmental impact is tracked using CodeCarbon, measuring:
54
  - Carbon emissions during inference
55
  - Energy consumption during inference
56
 
57
- This tracking helps establish a baseline for the environmental impact of model deployment and inference.
58
 
59
  ## Limitations
60
- - Makes completely random predictions
61
- - No learning or pattern recognition
62
- - No consideration of input text
63
- - Serves only as a baseline reference
64
- - Not suitable for any real-world applications
65
 
66
  ## Ethical Considerations
67
 
68
  - Dataset contains sensitive topics related to climate disinformation
69
- - Model makes random predictions and should not be used for actual classification
70
  - Environmental impact is tracked to promote awareness of AI's carbon footprint
71
  ```
 
8
  ---
9
 
10
 
11
+ # Fine-tuned ELECTRA model for Climate Disinformation Classification
12
 
13
  ## Model Description
14
 
15
+ This is our best-performing model for the Frugal AI Challenge 2024, specifically for the text classification task of identifying climate disinformation.
16
 
17
  ### Intended Use
18
 
19
+ - **Primary intended uses**: Comparison to baseline (random selection of labels) for climate disinformation classification models
20
  - **Primary intended users**: Researchers and developers participating in the Frugal AI Challenge
21
  - **Out-of-scope use cases**: Not intended for production use or real-world classification tasks
22
 
23
  ## Training Data
24
 
25
+ The model uses a balanced version of the QuotaClimat/frugalaichallenge-text-train training dataset. The dataset originally had the following structure:
26
  - Size: ~6000 examples
27
+ - Split: 80\% train, 20\% test
28
  - 8 categories of climate disinformation claims
29
 
30
  ### Labels
 
37
  6. Proponents are biased
38
  7. Fossil fuels are needed
39
 
40
+ The balancing was done to improve accuracy. We used the Marian MT model to augment our dataset by translating sentences from the classes with the lowest sample count to Spanish and back-translating them to English. The goal of this strategy was to keep sentences with similar sentences, but with different wording, under the same label. However, to avoid the dataset having more synthetic than original data, the target number of sentences per category was 2 times the smallest category. After augmenting the dataset, we removed duplicate sentences generated from back-translation, in order to avoid pseudoreplication. We then split the dataset into training and test sets, making sure to keep original and backtranslated within the same dataset to avoid data leakage.
41
+
42
+
43
  ## Performance
44
 
45
  ### Metrics
46
+ - **Accuracy**: Training - 0.91 , Validation - 0.87, Testing -
47
  - **Environmental Impact**:
48
  - Emissions tracked in gCO2eq
49
  - Energy consumption tracked in Wh
50
 
51
  ### Model Architecture
52
+ We fine-tuned a pre-trained ELECTRA model on our balanced dataset for five epochs. The first four layers of the model were frozen, while the training was carried out on the last eight layers. The Electra tokenizer was used to tokenize the sentences, with truncation set to True and padding to max length. An Adam optimizer was used, with a learning rate = 5e-5, epsilon=1e-7, beta 1=0.9 and beta_2=0.999. Our loss was Sparse Categorical Cross-entropy, and the main metric used was accuracy.
53
 
54
  ## Environmental Impact
55
 
 
57
  - Carbon emissions during inference
58
  - Energy consumption during inference
59
 
 
60
 
61
  ## Limitations
62
+
63
+ - The dataset was small to begin with, and every after augmenting the data, using a Neural Network on the dataset may lead to overfitting.
64
+ - While visual inspection of some sample augmented sentences suggested that the MariamMT model was successful in its back-translation, validation by subject matter experts is needed to guarantee that the augmented sentences maintain the label they were automatically assigned from the original sentence.
65
+
 
66
 
67
  ## Ethical Considerations
68
 
69
  - Dataset contains sensitive topics related to climate disinformation
 
70
  - Environmental impact is tracked to promote awareness of AI's carbon footprint
71
  ```