Upload train_qwen_codeforces.py with huggingface_hub
Browse files- train_qwen_codeforces.py +5 -1
train_qwen_codeforces.py
CHANGED
|
@@ -43,6 +43,10 @@ tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen3-0.6B")
|
|
| 43 |
if tokenizer.pad_token is None:
|
| 44 |
tokenizer.pad_token = tokenizer.eos_token
|
| 45 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
# Training configuration
|
| 47 |
config = SFTConfig(
|
| 48 |
# Hub settings - CRITICAL for saving results
|
|
@@ -94,11 +98,11 @@ peft_config = LoraConfig(
|
|
| 94 |
print("Initializing trainer...")
|
| 95 |
trainer = SFTTrainer(
|
| 96 |
model="Qwen/Qwen3-0.6B",
|
| 97 |
-
tokenizer=tokenizer,
|
| 98 |
train_dataset=train_dataset,
|
| 99 |
eval_dataset=eval_dataset,
|
| 100 |
args=config,
|
| 101 |
peft_config=peft_config,
|
|
|
|
| 102 |
)
|
| 103 |
|
| 104 |
print("Starting training...")
|
|
|
|
| 43 |
if tokenizer.pad_token is None:
|
| 44 |
tokenizer.pad_token = tokenizer.eos_token
|
| 45 |
|
| 46 |
+
# Processing function to convert messages to text using chat template
|
| 47 |
+
def formatting_func(example):
|
| 48 |
+
return tokenizer.apply_chat_template(example["messages"], tokenize=False)
|
| 49 |
+
|
| 50 |
# Training configuration
|
| 51 |
config = SFTConfig(
|
| 52 |
# Hub settings - CRITICAL for saving results
|
|
|
|
| 98 |
print("Initializing trainer...")
|
| 99 |
trainer = SFTTrainer(
|
| 100 |
model="Qwen/Qwen3-0.6B",
|
|
|
|
| 101 |
train_dataset=train_dataset,
|
| 102 |
eval_dataset=eval_dataset,
|
| 103 |
args=config,
|
| 104 |
peft_config=peft_config,
|
| 105 |
+
formatting_func=formatting_func,
|
| 106 |
)
|
| 107 |
|
| 108 |
print("Starting training...")
|