lrauch's picture
Update README.md
83339c7 verified
---
library_name: transformers
tags: []
---
# ConvNext (trained on XCL from BirdSet)
ConvNext trained on the XCL dataset from BirdSet, covering 9736 bird species from Xeno-Canto. Please refer to the [BirdSet Paper](https://arxiv.org/pdf/2403.10380) and the
[BirdSet Repository](https://github.com/DBD-research-group/BirdSet/tree/main) for further information.
### Model Details
ConvNeXT is a pure convolutional model (ConvNet), inspired by the design of Vision Transformers, that claims to outperform them.
## How to use
The BirdSet data needs a custom processor that is available in the BirdSet repository. The model does not have a processor available.
The model accepts a mono image (spectrogram) as input (e.g., `torch.Size([16, 1, 128, 343])`)
- The model is trained on 5-second clips of bird vocalizations.
- num_channels: 1
- pretrained checkpoint: facebook/convnext-base-224-22k
- sampling_rate: 32_000
- normalize spectrogram: mean: -4.268, std: 4.569 (from esc-50)
- spectrogram: n_fft: 1024, hop_length: 320, power: 2.0
- melscale: n_mels: 128, n_stft: 513
- dbscale: top_db: 80
```python
import torch
from transformers import AutoModelForImageClassification
from datasets import load_dataset
dataset = load_dataset("DBD-research-group/BirdSet", "HSN")
```
## Model Source
- **Repository:** [BirdSet Repository](https://github.com/DBD-research-group/BirdSet/tree/main)
- **Paper [optional]:** [BirdSet Paper](https://arxiv.org/pdf/2403.10380)
## Citation