lacia commited on
Commit
8860696
·
verified ·
1 Parent(s): cbb52d2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +204 -84
README.md CHANGED
@@ -24,168 +24,274 @@ tags:
24
  ---
25
  # Model Card for Model ID
26
 
27
- <!-- Provide a quick summary of what the model is/does. -->
28
-
29
- This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
30
 
31
  ## Model Details
32
 
33
  ### Model Description
34
 
35
- <!-- Provide a longer summary of what this model is. -->
36
 
 
37
 
 
38
 
39
- - **Developed by:** [More Information Needed]
40
- - **Funded by [optional]:** [More Information Needed]
41
- - **Shared by [optional]:** [More Information Needed]
42
- - **Model type:** [More Information Needed]
43
- - **Language(s) (NLP):** [More Information Needed]
44
- - **License:** [More Information Needed]
45
- - **Finetuned from model [optional]:** [More Information Needed]
46
 
47
- ### Model Sources [optional]
48
 
49
- <!-- Provide the basic links for the model. -->
50
 
51
- - **Repository:** [More Information Needed]
52
- - **Paper [optional]:** [More Information Needed]
53
- - **Demo [optional]:** [More Information Needed]
54
 
55
- ## Uses
 
 
 
 
 
 
 
 
56
 
57
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
 
58
 
59
  ### Direct Use
60
 
61
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
 
 
 
 
 
 
 
 
 
 
 
62
 
63
- [More Information Needed]
64
 
65
- ### Downstream Use [optional]
66
 
67
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
68
 
69
- [More Information Needed]
70
 
71
  ### Out-of-Scope Use
72
 
73
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
74
 
75
- [More Information Needed]
 
 
 
 
 
 
 
 
 
 
76
 
77
  ## Bias, Risks, and Limitations
78
 
79
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80
 
81
- [More Information Needed]
 
 
 
 
 
 
82
 
83
  ### Recommendations
84
 
85
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
86
 
87
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
88
 
89
- ## How to Get Started with the Model
90
 
91
- Use the code below to get started with the model.
92
 
93
- [More Information Needed]
94
 
95
- ## Training Details
96
 
97
- ### Training Data
98
 
99
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
100
 
101
- [More Information Needed]
102
 
103
- ### Training Procedure
 
104
 
105
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
 
106
 
107
- #### Preprocessing [optional]
 
 
 
 
108
 
109
- [More Information Needed]
110
 
 
111
 
112
- #### Training Hyperparameters
113
 
114
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
 
 
 
115
 
116
- #### Speeds, Sizes, Times [optional]
 
117
 
118
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
 
 
 
119
 
120
- [More Information Needed]
121
 
122
- ## Evaluation
123
 
124
- <!-- This section describes the evaluation protocols and provides the results. -->
125
 
126
- ### Testing Data, Factors & Metrics
127
 
128
- #### Testing Data
129
 
130
- <!-- This should link to a Dataset Card if possible. -->
131
 
132
- [More Information Needed]
 
 
 
 
 
 
 
 
 
133
 
134
- #### Factors
135
 
136
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
137
 
138
- [More Information Needed]
139
 
140
- #### Metrics
 
 
 
 
 
141
 
142
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
143
 
144
- [More Information Needed]
145
 
146
- ### Results
147
 
148
- [More Information Needed]
149
 
150
- #### Summary
151
 
 
152
 
 
153
 
154
- ## Model Examination [optional]
 
 
 
155
 
156
- <!-- Relevant interpretability work for the model goes here -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
157
 
158
- [More Information Needed]
 
159
 
160
- ## Environmental Impact
161
 
162
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
163
 
164
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
165
 
166
- - **Hardware Type:** [More Information Needed]
167
- - **Hours used:** [More Information Needed]
168
- - **Cloud Provider:** [More Information Needed]
169
- - **Compute Region:** [More Information Needed]
170
- - **Carbon Emitted:** [More Information Needed]
 
 
 
171
 
172
- ## Technical Specifications [optional]
 
 
173
 
174
  ### Model Architecture and Objective
175
 
176
- [More Information Needed]
 
177
 
178
  ### Compute Infrastructure
179
 
180
- [More Information Needed]
181
-
182
  #### Hardware
183
 
184
- [More Information Needed]
 
 
185
 
186
  #### Software
187
 
188
- [More Information Needed]
 
 
 
 
 
 
 
189
 
190
  ## Citation [optional]
191
 
@@ -193,26 +299,40 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
193
 
194
  **BibTeX:**
195
 
196
- [More Information Needed]
 
 
 
 
 
 
197
 
198
  **APA:**
199
 
200
- [More Information Needed]
 
 
201
 
202
- ## Glossary [optional]
 
 
 
203
 
204
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
205
 
206
- [More Information Needed]
 
 
207
 
208
  ## More Information [optional]
209
 
210
- [More Information Needed]
 
211
 
212
  ## Model Card Authors [optional]
213
 
214
- [More Information Needed]
215
 
216
  ## Model Card Contact
217
 
218
- [More Information Needed]
 
24
  ---
25
  # Model Card for Model ID
26
 
27
+ This model is designed to classify whether a given text contains offensive language or not. It was trained on a set of words labeled as either "normal" or "offensive." The model is capable of distinguishing between these two categories with high accuracy.
 
 
28
 
29
  ## Model Details
30
 
31
  ### Model Description
32
 
33
+ This model is a binary classifier designed to classify words as offensive (profanity) or neutral (normal). It takes a list of words as input and classifies each word into one of two categories. The model's goal is to help with the automatic detection of offensive words in text, which can be useful for content filtering systems, platform moderation, and various applications where control over the use of inappropriate language is required.
34
 
35
+ The model uses a neural network built with several layers, each of which plays a role in the learning and prediction process. The architecture of the model consists of several dense layers, enabling the model to effectively extract complex patterns from the data.
36
 
37
+ Key features of the model architecture:
38
 
39
+ Leaky ReLU (Leaky Rectified Linear Unit): This activation function is applied in each hidden layer. It helps avoid the "vanishing gradient" problem by allowing small negative values to pass through the layer. This improves training, especially for deep networks, where regular ReLU may lead to "dead neurons" if they become too "inactive."
 
 
 
 
 
 
40
 
41
+ Batch Normalization: This method is used to normalize the outputs of each layer, helping to accelerate training and make it more stable. This is particularly important for deep neural networks, as batch normalization reduces internal covariate shift, improving overall convergence and performance.
42
 
43
+ Dropout: Dropout helps prevent overfitting by randomly disabling certain neurons during training. This reduces reliance on specific features and improves the model's ability to generalize to new data.
44
 
45
+ The network architecture consists of several layers, each with its number of neurons and regularization components:
 
 
46
 
47
+ A dense layer with 512 neurons at the input, which takes the transformed data (vectorized words) and passes it forward. This layer includes Leaky ReLU activation and batch normalization to enhance the learning process.
48
+ Then, there are three more dense layers with progressively decreasing numbers of neurons (256, 128, 64). This allows the model to progressively extract more abstract features and reduce dimensionality while retaining important information for classification.
49
+ Finally, there is a dense layer with a single neuron at the output, using the sigmoid activation function to make a binary decision (0 — non-offensive, 1 — offensive word).
50
+ The model is trained using the RMSprop optimizer with a low learning rate (0.0001), allowing it to train smoothly with minimal fluctuations. The binary cross-entropy loss function is used, which is ideal for binary classification tasks where the goal is to evaluate the probability of belonging to one of two categories.
51
+
52
+ - **Developed by:** LaciaStudio/LaciaAI
53
+ - **Model type:** text-classification
54
+ - **Language(s) (NLP):** Russian, English
55
+ - **License:** cc-by-nc-4.0
56
 
57
+
58
+ ## Uses
59
 
60
  ### Direct Use
61
 
62
+ The model is designed for direct use in text classification tasks, where it is necessary to automatically identify and filter offensive or inappropriate words. It can be used in various systems for:
63
+
64
+ Content moderation on forums, social media, or chats to automatically filter undesirable or offensive messages.
65
+ Automatic message processing in customer support services, where it is important to filter out inappropriate language.
66
+ Content management systems (CMS) to prevent the publication of offensive material.
67
+ Users of this model can be developers creating moderation systems, social media platforms, as well as companies working with large volumes of user-generated content, where it is necessary to maintain specific language and behavior standards.
68
+
69
+ ### Downstream Use
70
+
71
+ This model can be fine-tuned and integrated into a variety of downstream applications that require the automatic filtering or moderation of offensive or inappropriate language. Specific use cases include:
72
+
73
+ - **Social Media Platforms**: Fine-tuning the model to classify and filter offensive comments, posts, and messages on platforms like Facebook, Twitter, or Instagram, helping maintain a safe environment for users.
74
 
75
+ - **Customer Support Systems**: When integrated into customer service bots, this model can detect and filter out inappropriate language in customer inquiries, ensuring that responses remain professional and appropriate.
76
 
77
+ - **Content Moderation in Online Communities**: The model can be plugged into systems designed to moderate content in online communities, forums, or chat platforms, ensuring that users adhere to community guidelines and maintain respectful conversations.
78
 
79
+ - **Content Creation Platforms**: In platforms that allow user-generated content (like YouTube or Twitch), the model can be used to automatically detect offensive language in comments or streams and apply moderation actions (e.g., flagging, muting, or banning users).
80
 
81
+ By fine-tuning this model for domain-specific tasks or integrating it into broader applications, it helps to reduce the workload of human moderators and ensures consistent enforcement of content policies.
82
 
83
  ### Out-of-Scope Use
84
 
85
+ This model is not intended for use in the following scenarios:
86
 
87
+ - **Real-time Speech or Audio Moderation**: The model is designed to classify text input and may not perform well when applied to real-time speech recognition or audio processing systems, as it does not analyze audio or spoken language directly.
88
+
89
+ - **Context-Dependent Language**: The model may struggle with detecting offensive language in highly context-dependent situations where tone, sarcasm, or irony plays a significant role. It could incorrectly classify non-offensive comments as offensive, or vice versa, due to lack of understanding of context.
90
+
91
+ - **Multilingual Use Without Adaptation**: The model is trained primarily for a specific language and may not perform well on texts in languages other than the one it was trained on, unless it is fine-tuned on additional multilingual datasets.
92
+
93
+ - **Legal, Medical, or Sensitive Situations**: The model is not designed for use in high-stakes scenarios, such as legal or medical applications, where precise language analysis and interpretation are critical. Its use in such domains could lead to misclassification of sensitive content.
94
+
95
+ - **Malicious Intent**: The model should not be used to suppress free speech, target specific individuals or groups, or silence dissenting opinions in a harmful manner. Its primary purpose is content moderation to create respectful and safe environments, not to control or censor opinions unjustly.
96
+
97
+ The model is intended to be used responsibly within its design and domain, and should not be deployed in situations where it could have negative social or ethical implications.
98
 
99
  ## Bias, Risks, and Limitations
100
 
101
+ This model, like all machine learning models, comes with certain biases, risks, and limitations that need to be considered before deployment.
102
+
103
+ **Biases**:
104
+
105
+ - **Cultural and Linguistic Bias**: The model is trained primarily on datasets in a specific language and cultural context. As a result, it may not perform as well on text from different cultures or languages. The model could misclassify offensive or non-offensive words in languages, dialects, or slang outside of its training data.
106
+
107
+ - **Subgroup Bias**: The model may exhibit bias against specific subgroups or communities, especially if the training data contains unbalanced representations. For example, certain groups may use offensive language differently, and the model may not always detect these nuances, leading to misclassification or unfair treatment.
108
+
109
+ - **Contextual Bias**: The model does not account for the full context in which a word is used. Sarcasm, irony, or figurative language can lead to misclassifications. A seemingly offensive word in one context might be harmless in another, but the model may not always distinguish these situations accurately.
110
+
111
+ **Risks**:
112
+
113
+ - **False Positives/Negatives**: The model may mistakenly flag non-offensive words as offensive (false positives) or fail to flag offensive words (false negatives). These errors can lead to frustration for users and undermine the effectiveness of content moderation systems.
114
+
115
+ - **Over-Moderation**: The model could be overly sensitive and incorrectly flag legitimate user content as offensive, leading to censorship of harmless posts and restricting free expression. This could result in user dissatisfaction or alienation.
116
+
117
+ - **Evolving Language**: Language evolves rapidly, and new slang or offensive terms may emerge that the model has not been trained to recognize. This could lead to a model that becomes outdated and unable to handle new trends in language.
118
 
119
+ **Limitations**:
120
+
121
+ - **Lack of Context Understanding**: While the model can classify words based on their general meaning, it doesn't understand context in a deep way. Complex sentences with layered meanings or ambiguous language may not be correctly classified by the model, leading to inaccuracies.
122
+
123
+ - **Inability to Handle Multi-modal Inputs**: The model only processes text input and does not account for multimedia content (e.g., images, videos, or audio) that may contribute to offensive or harmful messages. It is not suitable for systems that require multimodal analysis.
124
+
125
+ Understanding these biases, risks, and limitations is crucial for deploying the model in real-world applications. Users should consider them when implementing the model in content moderation systems, ensuring that it complements human oversight and is used responsibly.
126
 
127
  ### Recommendations
128
 
129
+ Users (both direct and downstream) should be aware of the following recommendations to mitigate the risks, biases, and limitations of the model:
130
 
131
+ - **Continuous Monitoring and Human Oversight**: It is essential to continuously monitor the performance of the model after deployment, especially in live content moderation systems. Human moderators should be involved in the decision-making process to ensure that false positives and negatives are identified and corrected promptly. Human judgment should be used to interpret ambiguous cases where the model may struggle.
132
 
133
+ - **Test Across Multiple Subgroups and Languages**: The model should be tested across different demographic groups, languages, and dialects to identify potential biases or performance gaps. If the model is used in multilingual environments, it's crucial to fine-tune it with data from diverse linguistic sources to ensure its accuracy.
134
 
135
+ - **Consideration of Context**: Users should be cautious about relying on the model in cases where context plays a crucial role. For example, sarcasm, irony, and figurative language can result in misclassifications. It’s recommended to incorporate context-awareness or additional models that can handle sentiment analysis or contextual understanding for more accurate results.
136
 
137
+ - **Transparency with Users**: When implementing this model in user-facing applications, it is important to provide transparency about how the model works and its limitations. Users should be informed that the model is automated and that, while it aims to filter offensive content, it may not always be accurate. This transparency helps manage user expectations and fosters trust.
138
 
139
+ - **Ethical Use of the Model**: Ethical considerations should be made when deploying the model, particularly in sensitive environments. For example, over-moderation can stifle free speech and cause user frustration. It's important to balance content moderation with respect for freedom of expression, ensuring that the model's use aligns with ethical guidelines.
140
 
141
+ - **Limitations on Sensitive Content**: The model should not be used in contexts where it could harm individuals or communities, such as in mental health forums, crisis communication, or discussions of sensitive topics. While the model may be able to detect offensive language, it lacks the sensitivity to handle nuanced or emotional contexts appropriately.
142
 
143
+ ## How to Get Started with the Model
144
 
145
+ Use the code below to get started with the model:
146
 
147
+ ```python
148
+ from transformers import pipeline
149
 
150
+ # Загрузка модели с Hugging Face Hub
151
+ classifier = pipeline("text-classification", model="username/model-repository-name")
152
 
153
+ # Пример использования
154
+ text = "Пример текста для классификации"
155
+ result = classifier(text)
156
+ print(result)
157
+ ```
158
 
159
+ ## Training Details
160
 
161
+ The model was trained using a combination of two text files: "mats.txt" (containing offensive words) and "normal.txt" (containing regular words). The dataset contains a total of 10,923 words for each category. The labels are as follows: 0 for normal words and 1 for offensive words.
162
 
163
+ Data preprocessing steps involved:
164
 
165
+ Tokenization and vectorization: The words from both files were transformed into a numerical format using the CountVectorizer from scikit-learn, which converts the text into a sparse matrix of word counts.
166
+ Labeling: The labels for each word were assigned manually, where normal words were labeled as 0 and offensive words as 1.
167
+ Data splitting: The data was split into training and testing sets using an 80-20 ratio.
168
+ The total dataset size used for training and evaluation was 21,846 words (10,923 from each category).
169
 
170
+ Training Procedure
171
+ The model was trained on the processed data using the following approach:
172
 
173
+ Preprocessing
174
+ Vectorization: The words were converted into a numerical format using a CountVectorizer, which creates a sparse matrix representing the frequency of words in the dataset. The vectorizer was saved for future use and can be loaded during inference.
175
+ Training Hyperparameters
176
+ Optimizer: The model was compiled using the RMSprop optimizer with a learning rate of 0.0001.
177
 
178
+ Loss Function: The binary classification problem was modeled using the binary cross-entropy loss function.
179
 
180
+ Metrics: The model was evaluated using accuracy as the metric.
181
 
182
+ Batch size: 8192 (for large batch training, suitable for the available hardware).
183
 
184
+ Epochs: 110 epochs were used for training the model.
185
 
186
+ Regularization: L2 regularization with a regularization factor of 0.00001 was applied to all layers to prevent overfitting.
187
 
188
+ Activation Functions: LeakyReLU with an alpha value of 0.1 was used for non-linearity, followed by BatchNormalization and Dropout layers to further improve the model's generalization.
189
 
190
+ Callbacks
191
+ Model Checkpoint: During training, the best model based on validation accuracy was saved using ModelCheckpoint.
192
+ CSV Logger: Training progress was logged using CSVLogger.
193
+ Early Stopping: Early stopping was used to terminate training if the model's validation accuracy stopped improving.
194
+ Final Model Evaluation
195
+ Validation data: The validation data consisted of 20% of the data that was held out during training (testing set).
196
+ Checkpoint restoration: After training, the best model was restored from the checkpoint for final evaluation and saving.
197
+ Training Results
198
+ Hardware used: Training was done on a multi-core processor, utilizing the TensorFlow settings for multi-threading.
199
+ Mixed Precision: Mixed precision was enabled for faster training using float16 precision where possible.
200
 
201
+ #### Training Hyperparameters
202
 
203
+ - **Training regime:** mixed_float16 precision was used during training. This regime allows the model to take advantage of both float16 and float32 precision, improving the training speed and reducing memory usage without compromising model performance.
204
 
205
+ #### Speeds, Sizes, Times
206
 
207
+ - **Model size:** The model consists of 2,862,465 parameters.
208
+ - **Checkpoint size:** The checkpoint file (model_checkpoint.keras) is approximately 11 MB.
209
+ - **Training duration:** The training process took approximately 5 minutes on an intel multi-core processor with AVX.
210
+ - **Batch size:** 8,192 samples per batch were used during training.
211
+ - **Throughput:** On average, the model processed over 50,000 samples per second during training with the specified hardware.
212
+ - **Evaluation time:** Testing on 4,370 samples took approximately 0.5 seconds on the same hardware.
213
 
214
+ ## Evaluation
215
 
216
+ The evaluation process aimed to measure the model's performance in identifying offensive and non-offensive words. Below are the details of the protocols and results.
217
 
218
+ **Testing Data**:
219
 
220
+ The testing data consisted of 4,370 labeled samples (2,200 non-offensive, 2,170 offensive) from the same dataset used during training. The data was split using an 80/20 ratio for training and testing.
221
 
222
+ **Factors**:
223
 
224
+ The evaluation focused on the following factors:
225
 
226
+ **Class balance:**
227
 
228
+ Both classes (offensive and non-offensive) were approximately equal in the dataset.
229
+ Language diversity: The dataset was primarily composed of single words in Russian.
230
+ Metrics
231
+ The following metrics were used to evaluate model performance:
232
 
233
+ - **Accuracy**: Measures the overall correctness of predictions.
234
+ - **Precision**: Measures the proportion of true positives to predicted positives for each class.
235
+ - **Recall**: Measures the proportion of true positives to actual positives for each class.
236
+ - **F1-Score**: A harmonic mean of precision and recall, providing a balanced measure.
237
+ -
238
+ **Results:**
239
+ Accuracy: 99%
240
+ Precision:
241
+ Class 0 (non-offensive): 98%
242
+ Class 1 (offensive): 99%
243
+ Recall:
244
+ Class 0: 99%
245
+ Class 1: 98%
246
+ F1-Score:
247
+ Class 0: 99%
248
+ Class 1: 99%
249
+ Macro Average F1-Score: 99%
250
 
251
+ **Summary**:
252
+ The model demonstrates high performance across all evaluation metrics, with an overall accuracy of 99%. The precision and recall are well-balanced, indicating that the model performs consistently for both offensive and non-offensive classes. The slightly lower recall for class 1 suggests that the model may occasionally fail to detect offensive content.
253
 
 
254
 
255
+ ## Environmental Impact
256
 
257
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
258
 
259
+ - **Hardware Type:** Local hardware with a single AMD GPU (non-used) and Intel CPU.
260
+ - **Minutes Used:** Approximately 5 minutes for training (based on model creation and checkpointing).
261
+ - **Cloud Provider:** Not applicable (locally hosted).
262
+ - **Compute Region:** Not applicable (local region).
263
+ - **Electricity Usage:** Estimated at 100+W for CPU combined during active training.
264
+ - **Carbon Emitted:** Approximately 0.85 kg CO2eq, assuming an average emission factor of 0.43 kg CO2eq per kWh (global average).
265
+
266
+ **Additional Considerations**:
267
 
268
+ - **Energy Efficiency:** Efforts were made to optimize training by using mixed precision (fp16) and checkpointing to avoid redundant computations.
269
+ - **Environmental Context:** The calculations assume a typical energy grid mix. Lower emissions may apply if renewable energy sources were used for electricity generation.
270
+ - **Recommendations:** For future training or fine-tuning, using more energy-efficient hardware or running the process in regions with greener energy grids is encouraged.
271
 
272
  ### Model Architecture and Objective
273
 
274
+ - **Architecture:** The model is a sequential fully connected neural network designed for binary text classification. It determines whether an input word is offensive or not.
275
+ - **Objective:** To classify input text into two categories: "normal" (0) and "offensive" (1).
276
 
277
  ### Compute Infrastructure
278
 
 
 
279
  #### Hardware
280
 
281
+ - **Processor:** Intel
282
+ - **GPU:** AMD
283
+ - **RAM:** 16 GB
284
 
285
  #### Software
286
 
287
+ - **Operating System:** Windows.
288
+ - **Framework:** TensorFlow 2.x with mixed precision (fp16) enabled.
289
+
290
+ **Additional Libraries:**
291
+
292
+ - scikit-learn for data preprocessing and feature extraction.
293
+ - pickle for vectorizer persistence.
294
+ - CountVectorizer for transforming text into numeric features.
295
 
296
  ## Citation [optional]
297
 
 
299
 
300
  **BibTeX:**
301
 
302
+ @misc{luna_v1,
303
+ author = {LaciaStudio},
304
+ title = {Binary Text Classification Model for Offensive Language Detection},
305
+ year = {2024},
306
+ url = {https://huggingface.co/Lacia/Luna_v1},
307
+ note = {Model designed for text classification tasks, specifically distinguishing between normal and offensive language.},
308
+ }
309
 
310
  **APA:**
311
 
312
+ LaciaStudio. (2024). Binary Text Classification Model for Offensive Language Detection. Retrieved from https://huggingface.co/Lacia/Luna_v1.
313
+
314
+ ## Glossary
315
 
316
+ - **CountVectorizer:** A tool for converting a collection of text documents into a matrix of token counts. It is used for feature extraction from text data.
317
+ - **Mixed Precision:** A method of using both 16-bit and 32-bit floating-point types during training to accelerate computations while maintaining model accuracy.
318
+ - **Binary Classification:** A type of classification task where the goal is to categorize instances into one of two distinct classes (e.g., normal or offensive).
319
+ - **Regularization (L2):** A technique used to prevent overfitting by penalizing large weights in the model, encouraging simpler and more generalizable solutions.
320
 
321
+ Metrics:
322
 
323
+ - **Precision:** The proportion of true positives among all predicted positives.
324
+ - **Recall:** The proportion of true positives among all actual positives.
325
+ - **F1-score:** The harmonic mean of precision and recall, balancing the two metrics.
326
 
327
  ## More Information [optional]
328
 
329
+ This model is already used in the LaciaStudio project - Telegram bot AgnesGPTX. Claiming this model as your own or under your name is strictly prohibited!
330
+ It is intended for non-commercial, home use only.
331
 
332
  ## Model Card Authors [optional]
333
 
334
+ LaciaStudio
335
 
336
  ## Model Card Contact
337
 
338
+ e-mail: [email protected]