|
--- |
|
title: Submission Oriaz |
|
emoji: 🔥 |
|
colorFrom: yellow |
|
colorTo: green |
|
sdk: docker |
|
pinned: True |
|
--- |
|
|
|
# Benchmarkusing different techniques |
|
|
|
## Global Informations : |
|
|
|
#### Intended Use |
|
|
|
- **Primary intended uses**: Baseline comparison for climate disinformation classification models |
|
- **Primary intended users**: Researchers and developers participating in the Frugal AI Challenge |
|
- **Out-of-scope use cases**: Not intended for production use or real-world classification tasks |
|
|
|
### Training Data |
|
|
|
The model uses the QuotaClimat/frugalaichallenge-text-train dataset: |
|
- Size: ~6000 examples |
|
- Split: 80% train, 20% test |
|
- 8 categories of climate disinformation claims |
|
|
|
#### Labels |
|
0. No relevant claim detected |
|
1. Global warming is not happening |
|
2. Not caused by humans |
|
3. Not bad or beneficial |
|
4. Solutions harmful/unnecessary |
|
5. Science is unreliable |
|
6. Proponents are biased |
|
7. Fossil fuels are needed |
|
|
|
### Environmental Impact |
|
|
|
Environmental impact is tracked using CodeCarbon, measuring: |
|
- Carbon emissions during inference |
|
- Energy consumption during inference |
|
|
|
This tracking helps establish a baseline for the environmental impact of model deployment and inference. |
|
|
|
### Ethical Considerations |
|
|
|
- Dataset contains sensitive topics related to climate disinformation |
|
- Environmental impact is tracked to promote awareness of AI's carbon footprint |
|
|
|
|
|
## ML model for Climate Disinformation Classification |
|
|
|
### Model Description |
|
|
|
Find the best ML model to process vectorized quotes to detect climate change disinformation. |
|
|
|
### Performance |
|
|
|
#### Metrics (I used NVIDIA T4 small GPU) |
|
- **Accuracy**: ~69-72% |
|
- **Environmental Impact**: |
|
- Emissions tracked in gCO2eq (~0,7g) |
|
- Energy consumption tracked in Wh (~1,8wh) |
|
|
|
#### Model Architecture |
|
|
|
ML models prefers numeric values so we need to embed our quotes. I used *MTEB Leaderboard* on HuggingFace to find the model with the best trade-off between performance and the number of parameters. |
|
|
|
I then chosed "dunzhang/stella_en_400M_v5" model as embedder. It has the 7th best performance score with only 400M parameters. |
|
|
|
Once the quote are embedded, I have 6091 values x 1024 features. After that, train-test split (70%, 30%). |
|
|
|
Using TPOT Classifier, I found that the best model on my data was a Logistic Regressor. |
|
|
|
Then here is the Confusion Matrix : |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/66169e1ce557753f30eab31b/tfAcfFu3Cnc9XJ00ixrWB.png) |
|
|
|
### Limitations |
|
- Embedding phase take ~30 secondes for 1800 quotes. It can be optimised and can have a real influence on carbon emissions. |
|
- Hard to go over 70% accuracy with "simple" ML. |
|
- Textual data have some interpretations limitations that little models can't find. |
|
|
|
|
|
|
|
## Bert model for Climate Disinformation Classification |
|
|
|
### Model Description |
|
|
|
Fine tune model for model classification. |
|
|
|
### Performance |
|
|
|
#### Metrics (I used NVIDIA T4 small GPU) |
|
- **Accuracy**: ~90% |
|
- **Environmental Impact**: |
|
- Emissions tracked in gCO2eq (~0,25g) |
|
- Energy consumption tracked in Wh (~0.7wh) |
|
|
|
#### Model Architecture |
|
|
|
Fine tuning of "bert-uncased" model with 70% train, 15% eval, 15% test datasets. |
|
|
|
### Limitations |
|
- Not optimized. I need to try to run it on CPU |
|
- Little models have limitations. Regularly between 70-80% accuracy. Hard to go over just by changing params. |
|
|
|
# Contacts : |
|
*LinkedIn* : Mattéo GIRARDEAU |
|
*email* : [email protected] |
|
``` |
|
|