Spaces:
Configuration error
Configuration error
license: apache-2.0 | |
language: | |
- en | |
tags: | |
- sft | |
pipeline_tag: text-generation | |
widget: | |
- text: <|prompter|>What is a meme, and what's the history behind this word?<|endoftext|><|assistant|> | |
- text: <|prompter|>What's the Earth total population<|endoftext|><|assistant|> | |
- text: <|prompter|>Write a story about future of AI development<|endoftext|><|assistant|> | |
# Open-Assistant SFT-4 12B Model | |
This is the 4th iteration English supervised-fine-tuning (SFT) model of | |
the [Open-Assistant](https://github.com/LAION-AI/Open-Assistant) project. | |
It is based on a Pythia 12B that was fine-tuned on human demonstrations | |
of assistant conversations collected through the | |
[https://open-assistant.io/](https://open-assistant.io/) human feedback web | |
app before March 25, 2023. | |
## Model Details | |
- **Developed by:** [Open-Assistant Contributors](https://open-assistant.io/) | |
- **Model type:** Transformer-based Language Model | |
- **Language:** English | |
- **Finetuned from:** [EleutherAI / pythia-12b-deduped](https://huggingface.co/EleutherAI/pythia-12b-deduped) | |
- **Code:** [Open-Assistant/model/model_training](https://github.com/LAION-AI/Open-Assistant/tree/main/model/model_training) | |
- **Demo:** [Continuations for 250 random prompts](https://open-assistant.github.io/oasst-model-eval/?f=https%3A%2F%2Fraw.githubusercontent.com%2FOpen-Assistant%2Foasst-model-eval%2Fmain%2Fsampling_reports%2Foasst-sft%2F2023-04-03_andreaskoepf_oasst-sft-4-pythia-12b-epoch-3_5_sampling_noprefix_lottery.json%0Ahttps%3A%2F%2Fraw.githubusercontent.com%2FOpen-Assistant%2Foasst-model-eval%2Fmain%2Fsampling_reports%2Fchat-gpt%2F2023-04-11_gpt-3.5-turbo_lottery.json) | |
- **License:** Apache 2.0 | |
- **Contact:** [Open-Assistant Discord](https://ykilcher.com/open-assistant-discord) | |
## Prompting | |
Two special tokens are used to mark the beginning of user and assistant turns: | |
`<|prompter|>` and `<|assistant|>`. Each turn ends with a `<|endoftext|>` token. | |
Input prompt example: | |
``` | |
<|prompter|>What is a meme, and what's the history behind this word?<|endoftext|><|assistant|> | |
``` | |
The input ends with the `<|assistant|>` token to signal that the model should | |
start generating the assistant reply. | |
## Dev Details | |
- wandb: https://wandb.ai/open-assistant/supervised-finetuning/runs/770a0t41 | |
- base model: [andreaskoepf/pythia-12b-pre-2000](https://huggingface.co/andreaskoepf/pythia-12b-pre-2000) | |
- checkpoint: 4000 steps | |
command: `deepspeed trainer_sft.py --configs defaults reference-data reference-pythia-12b --cache_dir /home/ubuntu/data_cache --output_dir .saved/oasst-sft-3-pythia-12b-reference_2kpre --num_train_epochs 8 --residual_dropout 0.2 --deepspeed --use_flash_attention true --model_name andreaskoepf/pythia-12b-pre-2000` | |
data: | |
``` | |
reference-data: | |
datasets: | |
- oasst_export: | |
lang: "bg,ca,cs,da,de,en,es,fr,hr,hu,it,nl,pl,pt,ro,ru,sl,sr,sv,uk" | |
input_file_path: 2023-03-25_oasst_research_ready_synth_labels.jsonl.gz | |
val_split: 0.05 | |
- alpaca | |
sort_by_length: false | |
use_custom_sampler: false | |
``` | |
pythia: | |
``` | |
reference-pythia-12b: | |
dtype: fp16 | |
log_dir: "pythia_log_12b" | |
learning_rate: 6e-6 | |
model_name: EleutherAI/pythia-12b-deduped | |
output_dir: pythia_model_12b | |
weight_decay: 0.0 | |
max_length: 2048 | |
warmup_steps: 100 | |
gradient_checkpointing: true | |
gradient_accumulation_steps: 2 | |
per_device_train_batch_size: 4 | |
per_device_eval_batch_size: 4 | |
eval_steps: 100 | |
save_steps: 1000 | |
num_train_epochs: 8 | |
save_total_limit: 4 | |
``` | |
zero config: | |
``` | |
{ | |
"fp16": { | |
"enabled": "auto", | |
"loss_scale": 0, | |
"loss_scale_window": 1000, | |
"initial_scale_power": 16, | |
"hysteresis": 2, | |
"min_loss_scale": 1 | |
}, | |
"bf16": { | |
"enabled": "auto" | |
}, | |
"optimizer": { | |
"type": "AdamW", | |
"params": { | |
"lr": "auto", | |
"betas": "auto", | |
"eps": "auto", | |
"weight_decay": "auto" | |
} | |
}, | |
"scheduler": { | |
"type": "WarmupDecayLR", | |
"params": { | |
"warmup_min_lr": "auto", | |
"warmup_max_lr": "auto", | |
"warmup_num_steps": "auto", | |
"total_num_steps": "auto" | |
} | |
}, | |
"zero_optimization": { | |
"stage": 2, | |
"allgather_partitions": true, | |
"allgather_bucket_size": 1e9, | |
"overlap_comm": false, | |
"reduce_scatter": true, | |
"reduce_bucket_size": 1e9, | |
"contiguous_gradients": true | |
}, | |
"gradient_accumulation_steps": "auto", | |
"gradient_clipping": "auto", | |
"steps_per_print": 2000, | |
"train_batch_size": "auto", | |
"train_micro_batch_size_per_gpu": "auto", | |
"wall_clock_breakdown": false | |
} | |
``` |