Spaces:
Running
Running
adds quantization configuration correctly
Browse files
scripts/training/train_gpt_oss.py
CHANGED
|
@@ -70,9 +70,9 @@ def setup_lora_for_gpt_oss(model, config):
|
|
| 70 |
|
| 71 |
# LoRA configuration as per tutorial
|
| 72 |
lora_config = LoraConfig(
|
| 73 |
-
r=config.lora_config.get("r", 8),
|
| 74 |
-
lora_alpha=config.lora_config.get("lora_alpha", 16),
|
| 75 |
-
target_modules=config.lora_config.get("target_modules", "all-linear"),
|
| 76 |
target_parameters=config.lora_config.get("target_parameters", [
|
| 77 |
"7.mlp.experts.gate_up_proj",
|
| 78 |
"7.mlp.experts.down_proj",
|
|
@@ -80,7 +80,14 @@ def setup_lora_for_gpt_oss(model, config):
|
|
| 80 |
"15.mlp.experts.down_proj",
|
| 81 |
"23.mlp.experts.gate_up_proj",
|
| 82 |
"23.mlp.experts.down_proj",
|
| 83 |
-
])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 84 |
)
|
| 85 |
|
| 86 |
peft_model = get_peft_model(model, lora_config)
|
|
|
|
| 70 |
|
| 71 |
# LoRA configuration as per tutorial
|
| 72 |
lora_config = LoraConfig(
|
| 73 |
+
r=config.lora_config.get("r", 8) if config.lora_config else 8,
|
| 74 |
+
lora_alpha=config.lora_config.get("lora_alpha", 16) if config.lora_config else 16,
|
| 75 |
+
target_modules=config.lora_config.get("target_modules", "all-linear") if config.lora_config else "all-linear",
|
| 76 |
target_parameters=config.lora_config.get("target_parameters", [
|
| 77 |
"7.mlp.experts.gate_up_proj",
|
| 78 |
"7.mlp.experts.down_proj",
|
|
|
|
| 80 |
"15.mlp.experts.down_proj",
|
| 81 |
"23.mlp.experts.gate_up_proj",
|
| 82 |
"23.mlp.experts.down_proj",
|
| 83 |
+
]) if config.lora_config else [
|
| 84 |
+
"7.mlp.experts.gate_up_proj",
|
| 85 |
+
"7.mlp.experts.down_proj",
|
| 86 |
+
"15.mlp.experts.gate_up_proj",
|
| 87 |
+
"15.mlp.experts.down_proj",
|
| 88 |
+
"23.mlp.experts.gate_up_proj",
|
| 89 |
+
"23.mlp.experts.down_proj",
|
| 90 |
+
],
|
| 91 |
)
|
| 92 |
|
| 93 |
peft_model = get_peft_model(model, lora_config)
|