YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

convert this to raw readme.md file, it's a model card on huggingface

Pashto BERT (BERT-Base)

Model Overview

This is a monolingual Pashto BERT (BERT-Base) model trained on a large Pashto corpus. The model is designed to understand and generate text in Pashto, making it suitable for various downstream Natural Language Processing (NLP) tasks.

Model Details

  • Architecture: BERT-Base (12 layers, 768 hidden size, 12 attention heads, 110M parameters)
  • Language: Pashto (ps)
  • Training Corpus: A diverse set of Pashto text data, including news articles, books, and web content.
  • Special Tokens: [CLS], [SEP], [PAD], [MASK], [UNK]

Intended Use

This model can be fine-tuned for various Pashto-specific NLP tasks, such as:

  • Sequence Classification: Sentiment analysis, topic classification, and document categorization.
  • Sequence Tagging: Named entity recognition (NER) and part-of-speech (POS) tagging.
  • Text Generation & Understanding: Question answering, text summarization, and machine translation.

How to Use

This model can be loaded using the transformers library from Hugging Face:

from transformers import AutoModel, AutoTokenizer

model_name = "your-huggingface-username/pashto-bert-base"
tokenizer = AutoTokenizer.from_pretrained("/kaggle/working/model/")
model = AutoModel.from_pretrained(model_name)

text = "ستاسو نننۍ ورځ څنګه وه؟"
tokens = tokenizer(text, return_tensors="pt")
out = model(**tokens)

Training Details

  • Optimization: AdamW
  • Sequence Length: 128
  • Warmup Steps: 10,000
  • Warmup Ratio: 0.06
  • Learning Rate: 1e-4
  • Weight Decay: 0.01
  • Adam Optimizer Parameters:
    • Epsilon: 1e-8
    • Betas: (0.9, 0.999)
  • Gradient Accumulation Steps: 1
  • Max Gradient Norm: 1.0
  • Scheduler: linear_schedule_with_warmup

Limitations & Biases

  • The model may reflect biases present in the training data.
  • Performance on low-resource or domain-specific tasks may require additional fine-tuning.
  • It is not trained for code-switching scenarios (e.g., mixing Pashto with English or other languages).
Downloads last month
0
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Collection including zirak-ai/pashto-bert-v1