cpt core 4
Browse files
scripts/cpt_core_model_4.py
CHANGED
@@ -41,7 +41,8 @@ model = FastLanguageModel.get_peft_model(
|
|
41 |
lora_dropout = 0, # Supports any, but = 0 is optimized
|
42 |
bias = "none", # Supports any, but = "none" is optimized
|
43 |
# [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
|
44 |
-
use_gradient_checkpointing = "unsloth", # True or "unsloth" for very long context
|
|
|
45 |
random_state = 3407,
|
46 |
use_rslora = True, # We support rank stabilized LoRA
|
47 |
loftq_config = None, # And LoftQ
|
|
|
41 |
lora_dropout = 0, # Supports any, but = 0 is optimized
|
42 |
bias = "none", # Supports any, but = "none" is optimized
|
43 |
# [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
|
44 |
+
# use_gradient_checkpointing = "unsloth", # True or "unsloth" for very long context
|
45 |
+
use_gradient_checkpointing = False,
|
46 |
random_state = 3407,
|
47 |
use_rslora = True, # We support rank stabilized LoRA
|
48 |
loftq_config = None, # And LoftQ
|