qgallouedec HF Staff commited on
Commit
4d346dc
·
1 Parent(s): 728c826
configs/Qwen3-4B-Base_a10g-small.yaml CHANGED
@@ -1,2 +1,44 @@
 
1
  model_name_or_path: Qwen/Qwen2.5-0.5B
2
- dataset_name: stanfordnlp/imdb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model arguments
2
  model_name_or_path: Qwen/Qwen2.5-0.5B
3
+ model_revision: main
4
+ torch_dtype: bfloat16
5
+ attn_implementation: flash_attention_2
6
+
7
+ # Data training arguments
8
+ dataset_name: trl-lib/tldr
9
+ dataset_num_proc: 4
10
+
11
+ # SFT trainer config
12
+ bf16: true
13
+ do_eval: false
14
+ eval_strategy: 'no'
15
+ gradient_accumulation_steps: 8
16
+ gradient_checkpointing: true
17
+ gradient_checkpointing_kwargs:
18
+ use_reentrant: false
19
+ # hub_model_id: open-r1/OlympicCoder-7B
20
+ hub_strategy: every_save
21
+ learning_rate: 1.0e-05
22
+ log_level: info
23
+ logging_steps: 1
24
+ logging_strategy: steps
25
+ lr_scheduler_type: cosine_with_min_lr
26
+ lr_scheduler_kwargs:
27
+ min_lr_rate: 0.1
28
+ packing: false
29
+ max_grad_norm: 0.2
30
+ max_length: 512
31
+ max_steps: -1
32
+ num_train_epochs: 10
33
+ output_dir: data/Qwen2.5-0.5B-SFT
34
+ overwrite_output_dir: true
35
+ per_device_eval_batch_size: 1
36
+ per_device_train_batch_size: 2
37
+ push_to_hub: true
38
+ report_to:
39
+ - wandb
40
+ save_strategy: epoch
41
+ save_total_limit: 1
42
+ seed: 42
43
+ use_liger_kernel: true
44
+ warmup_ratio: 0.03