cpt core 4
Browse files
scripts/cpt_core_model_4.py
CHANGED
@@ -34,16 +34,16 @@ model, tokenizer = FastLanguageModel.from_pretrained(
|
|
34 |
|
35 |
model = FastLanguageModel.get_peft_model(
|
36 |
model,
|
37 |
-
|
38 |
-
r=16, # Choose any number > 0 ! Suggested 8, 16, 32, 64, 128
|
39 |
target_modules=[
|
40 |
"q_proj", "k_proj", "v_proj", "o_proj",
|
41 |
"gate_proj",
|
42 |
"up_proj", "down_proj",
|
43 |
"embed_tokens", "lm_head",
|
44 |
],
|
45 |
-
|
46 |
-
lora_alpha=16,
|
47 |
lora_dropout=0, # Supports any, but = 0 is optimized
|
48 |
bias='none', # Supports any, but = "none" is optimized
|
49 |
# [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
|
|
|
34 |
|
35 |
model = FastLanguageModel.get_peft_model(
|
36 |
model,
|
37 |
+
r=256, # Choose any number > 0 ! Suggested 8, 16, 32, 64, 128
|
38 |
+
# r=16, # Choose any number > 0 ! Suggested 8, 16, 32, 64, 128
|
39 |
target_modules=[
|
40 |
"q_proj", "k_proj", "v_proj", "o_proj",
|
41 |
"gate_proj",
|
42 |
"up_proj", "down_proj",
|
43 |
"embed_tokens", "lm_head",
|
44 |
],
|
45 |
+
lora_alpha=32,
|
46 |
+
# lora_alpha=16,
|
47 |
lora_dropout=0, # Supports any, but = 0 is optimized
|
48 |
bias='none', # Supports any, but = "none" is optimized
|
49 |
# [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
|