File size: 1,607 Bytes
dcf97f6 835282c dcf97f6 7c2716e dcf97f6 61c4e78 e372aa0 3d051ef e372aa0 5fb9b8a 0f751d1 3d051ef dcf97f6 27dd81f 8480c67 27dd81f dcf97f6 7ef0ed0 dcf97f6 7ef0ed0 dcf97f6 835282c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 |
---
license: cc-by-nc-2.0
base_model: facebook/opt-350m
tags:
- generated_from_trainer
model-index:
- name: tmp_trainer
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# tmp_trainer
This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m) on the [addressWithContext](https://huggingface.co/datasets/piazzola/addressWithContext) dataset.
## Model description
**Make sure to set max_new_tokens = 20; otherwise, the model will generate one token at a time.**
```
nlp = pipeline("text-generation",
model="piazzola/tmp_trainer",
max_new_tokens=20)
nlp("I live at 15 Firstfield Road.")
```
**Note that if you would like to try longer sentences using the Hosted inference API
on the right hand side on this website, you might need to click "Compute" more than one time to get the address.**
## Intended uses & limitations
The model is intended to detect addresses that occur in a sentence.
## Training and evaluation data
This model is trained on `piazzola/addressWithContext`.
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Framework versions
- Transformers 4.34.0
- Pytorch 2.0.1+cu118
- Datasets 2.14.5
- Tokenizers 0.14.1 |