Spaces:
Paused
Paused
Update finetune.py
Browse files- finetune.py +4 -7
finetune.py
CHANGED
@@ -17,12 +17,12 @@ from peft import (
|
|
17 |
|
18 |
|
19 |
# Parameters
|
20 |
-
MICRO_BATCH_SIZE = int(
|
21 |
BATCH_SIZE = 16
|
22 |
-
size =
|
23 |
GRADIENT_ACCUMULATION_STEPS = BATCH_SIZE // MICRO_BATCH_SIZE
|
24 |
EPOCHS = 1
|
25 |
-
LEARNING_RATE = float(
|
26 |
CUTOFF_LEN = 512
|
27 |
LORA_R = 8
|
28 |
LORA_ALPHA = 16
|
@@ -43,7 +43,7 @@ if not os.path.exists("data"):
|
|
43 |
os.makedirs("data")
|
44 |
# Load data
|
45 |
data = []
|
46 |
-
for x in
|
47 |
data += json.load(open("data/{}_chat_data.json".format(x)))
|
48 |
random.shuffle(data)
|
49 |
json.dump(data, open(DATA_PATH, "w"))
|
@@ -161,9 +161,6 @@ model.state_dict = (
|
|
161 |
lambda self, *_, **__: get_peft_model_state_dict(self, old_state_dict())
|
162 |
).__get__(model, type(model))
|
163 |
|
164 |
-
if torch.__version__ >= "2" and sys.platform != "win32":
|
165 |
-
model = torch.compile(model)
|
166 |
-
|
167 |
trainer.train()
|
168 |
|
169 |
model.save_pretrained(OUTPUT_DIR)
|
|
|
17 |
|
18 |
|
19 |
# Parameters
|
20 |
+
MICRO_BATCH_SIZE = int(32)
|
21 |
BATCH_SIZE = 16
|
22 |
+
size = 7b
|
23 |
GRADIENT_ACCUMULATION_STEPS = BATCH_SIZE // MICRO_BATCH_SIZE
|
24 |
EPOCHS = 1
|
25 |
+
LEARNING_RATE = float(0.00015)
|
26 |
CUTOFF_LEN = 512
|
27 |
LORA_R = 8
|
28 |
LORA_ALPHA = 16
|
|
|
43 |
os.makedirs("data")
|
44 |
# Load data
|
45 |
data = []
|
46 |
+
for x in "alpaca,stackoverflow,quora".split(","):
|
47 |
data += json.load(open("data/{}_chat_data.json".format(x)))
|
48 |
random.shuffle(data)
|
49 |
json.dump(data, open(DATA_PATH, "w"))
|
|
|
161 |
lambda self, *_, **__: get_peft_model_state_dict(self, old_state_dict())
|
162 |
).__get__(model, type(model))
|
163 |
|
|
|
|
|
|
|
164 |
trainer.train()
|
165 |
|
166 |
model.save_pretrained(OUTPUT_DIR)
|