data_cfgs: eval_data_files: null eval_datasets: null eval_optional_args: [] eval_size: null eval_split: null eval_subset: null eval_template: null train_data_files: llf_ti2ti_13.5k_tokenized.pt train_datasets: /data/align-anything/hantao/align-anything/projects/text_image_to_text_image/outputs train_optional_args: [] train_size: null train_split: train train_subset: null train_template: AA_textfeedback logger_cfgs: cache_dir: null log_project: align-anything log_run_name: sft log_type: wandb output_dir: ../outputs/sft_tf_cham_1111_13.5k_ti2ti save_interval: 400.0 model_cfgs: model_max_length: 4096 model_name_or_path: /data/align-anything/hantao/models/0916_ti_to_ti_sft/ trust_remote_code: true special_tokens: null train_cfgs: adam_betas: - 0.9 - 0.95 adam_epsilon: 1.0e-08 bf16: true ds_cfgs: ds_z3_config.json epochs: 3.0 eval_interval: 1000 eval_strategy: steps fp16: false freeze_language_model: false gradient_accumulation_steps: 2.0 gradient_checkpointing: true learning_rate: 1.0e-06 lr_scheduler_type: cosine lr_warmup_ratio: 0.03 max_grad_norm: 1.0 per_device_eval_batch_size: 4.0 per_device_train_batch_size: 4.0 seed: 42 weight_decay: 0.0