ansi-code's picture
Update README.md
ac66fca
---
license: apache-2.0
language:
- en
metrics:
- accuracy
pipeline_tag: text-classification
tags:
- advertising
---
# Tiny Bert Domain Advertising Classifier
https://huggingface.co/ansi-code/bert-domain-advertising-classifier/blob/main/bert_domain_advertising_classifier.ipynb
## Overview
AdTargetingBERTClassifier is a small-scale BERT-based classifier designed for the task of ad targeting classification. The model is trained to predict multi-class labels associated with domains, as provided in the DAC693K dataset.
## Model Architecture
The classifier is built on the BERT (Bidirectional Encoder Representations from Transformers) architecture. It takes domain text as input and outputs logits for each class, enabling multi-class classification for ad targeting.
## Model Training
The model is trained on the "AdTargetingDataset" using a supervised learning approach. The training involves optimizing for the categorical cross-entropy loss, and the model is fine-tuned on the specific ad targeting classes associated with each domain.
## Usage
### Loading the Model
To use the trained classifier in your Python environment, you can load it using the following code:
```python
from transformers import BertTokenizer, BertForSequenceClassification
import torch
# Load the pre-trained model and tokenizer
model = BertForSequenceClassification.from_pretrained("ansi-code/bert-domain-advertising-classifier")
tokenizer = BertTokenizer.from_pretrained("bert-base-uncased")
# Example inference
text = "google.com"
inputs = tokenizer(text, return_tensors="pt")
outputs = model(**inputs)
logits = outputs.logits
```
## Prediction
To make predictions with the loaded model, you can use the obtained logits. Convert the logits to probabilities and determine the predicted class based on the highest probability.
```python
Copy code
probabilities = torch.nn.functional.sigmoid(logits, dim=-1)
predicted_class = torch.argmax(probabilities).item()
```
## Model Evaluation
The model's performance can be assessed using standard evaluation metrics such as accuracy, precision, recall, and F1-score on a separate validation set or through cross-validation.
## License
This model is released under the Apache 2.0 License.
## Citation
If you use this model in your work, please cite it using the following BibTeX entry:
```bibtex
@model{silvi_2023_bert-domain-advertising-classifier,
title = {bert-domain-advertising-classifier},
author = {Andrea Silvi},
year = {2023},
}
```
## Acknowledgements
We would like to thank the developers of the Hugging Face Transformers library for providing the BERT model implementation.