|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
metrics: |
|
- accuracy |
|
pipeline_tag: text-classification |
|
tags: |
|
- advertising |
|
--- |
|
|
|
# Tiny Bert Domain Advertising Classifier |
|
|
|
https://huggingface.co/ansi-code/bert-domain-advertising-classifier/blob/main/bert_domain_advertising_classifier.ipynb |
|
|
|
## Overview |
|
|
|
AdTargetingBERTClassifier is a small-scale BERT-based classifier designed for the task of ad targeting classification. The model is trained to predict multi-class labels associated with domains, as provided in the DAC693K dataset. |
|
|
|
## Model Architecture |
|
|
|
The classifier is built on the BERT (Bidirectional Encoder Representations from Transformers) architecture. It takes domain text as input and outputs logits for each class, enabling multi-class classification for ad targeting. |
|
|
|
## Model Training |
|
|
|
The model is trained on the "AdTargetingDataset" using a supervised learning approach. The training involves optimizing for the categorical cross-entropy loss, and the model is fine-tuned on the specific ad targeting classes associated with each domain. |
|
|
|
## Usage |
|
|
|
### Loading the Model |
|
|
|
To use the trained classifier in your Python environment, you can load it using the following code: |
|
|
|
```python |
|
from transformers import BertTokenizer, BertForSequenceClassification |
|
import torch |
|
|
|
# Load the pre-trained model and tokenizer |
|
model = BertForSequenceClassification.from_pretrained("ansi-code/bert-domain-advertising-classifier") |
|
tokenizer = BertTokenizer.from_pretrained("bert-base-uncased") |
|
|
|
# Example inference |
|
text = "google.com" |
|
inputs = tokenizer(text, return_tensors="pt") |
|
outputs = model(**inputs) |
|
logits = outputs.logits |
|
``` |
|
|
|
## Prediction |
|
To make predictions with the loaded model, you can use the obtained logits. Convert the logits to probabilities and determine the predicted class based on the highest probability. |
|
|
|
```python |
|
Copy code |
|
probabilities = torch.nn.functional.sigmoid(logits, dim=-1) |
|
predicted_class = torch.argmax(probabilities).item() |
|
``` |
|
|
|
## Model Evaluation |
|
|
|
The model's performance can be assessed using standard evaluation metrics such as accuracy, precision, recall, and F1-score on a separate validation set or through cross-validation. |
|
|
|
## License |
|
|
|
This model is released under the Apache 2.0 License. |
|
|
|
## Citation |
|
|
|
If you use this model in your work, please cite it using the following BibTeX entry: |
|
|
|
```bibtex |
|
@model{silvi_2023_bert-domain-advertising-classifier, |
|
title = {bert-domain-advertising-classifier}, |
|
author = {Andrea Silvi}, |
|
year = {2023}, |
|
} |
|
``` |
|
|
|
## Acknowledgements |
|
|
|
We would like to thank the developers of the Hugging Face Transformers library for providing the BERT model implementation. |
|
|