gustavoaq commited on
Commit
bbf226a
·
1 Parent(s): 92fa0e7

Update finetune.py

Browse files
Files changed (1) hide show
  1. finetune.py +4 -7
finetune.py CHANGED
@@ -17,12 +17,12 @@ from peft import (
17
 
18
 
19
  # Parameters
20
- MICRO_BATCH_SIZE = int(sys.argv[2])
21
  BATCH_SIZE = 16
22
- size = sys.argv[1]
23
  GRADIENT_ACCUMULATION_STEPS = BATCH_SIZE // MICRO_BATCH_SIZE
24
  EPOCHS = 1
25
- LEARNING_RATE = float(sys.argv[3])
26
  CUTOFF_LEN = 512
27
  LORA_R = 8
28
  LORA_ALPHA = 16
@@ -43,7 +43,7 @@ if not os.path.exists("data"):
43
  os.makedirs("data")
44
  # Load data
45
  data = []
46
- for x in sys.argv[4].split(","):
47
  data += json.load(open("data/{}_chat_data.json".format(x)))
48
  random.shuffle(data)
49
  json.dump(data, open(DATA_PATH, "w"))
@@ -161,9 +161,6 @@ model.state_dict = (
161
  lambda self, *_, **__: get_peft_model_state_dict(self, old_state_dict())
162
  ).__get__(model, type(model))
163
 
164
- if torch.__version__ >= "2" and sys.platform != "win32":
165
- model = torch.compile(model)
166
-
167
  trainer.train()
168
 
169
  model.save_pretrained(OUTPUT_DIR)
 
17
 
18
 
19
  # Parameters
20
+ MICRO_BATCH_SIZE = int(32)
21
  BATCH_SIZE = 16
22
+ size = 7b
23
  GRADIENT_ACCUMULATION_STEPS = BATCH_SIZE // MICRO_BATCH_SIZE
24
  EPOCHS = 1
25
+ LEARNING_RATE = float(0.00015)
26
  CUTOFF_LEN = 512
27
  LORA_R = 8
28
  LORA_ALPHA = 16
 
43
  os.makedirs("data")
44
  # Load data
45
  data = []
46
+ for x in "alpaca,stackoverflow,quora".split(","):
47
  data += json.load(open("data/{}_chat_data.json".format(x)))
48
  random.shuffle(data)
49
  json.dump(data, open(DATA_PATH, "w"))
 
161
  lambda self, *_, **__: get_peft_model_state_dict(self, old_state_dict())
162
  ).__get__(model, type(model))
163
 
 
 
 
164
  trainer.train()
165
 
166
  model.save_pretrained(OUTPUT_DIR)