nroggendorff commited on
Commit
6f9e2c0
·
verified ·
1 Parent(s): d9cef8e

Update train.py

Browse files
Files changed (1) hide show
  1. train.py +3 -3
train.py CHANGED
@@ -3,11 +3,11 @@ import os
3
  import torch
4
  import trl
5
 
6
- from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM, TrainingArguments, PreTrainedTokenizerFast, AdamW, get_cosine_schedule_with_warmup
7
  from datasets import load_dataset, Dataset
8
  from tokenizers import ByteLevelBPETokenizer
9
 
10
- BATCH_SIZE = 1
11
  EPOCHS = 2
12
  LEARNING_RATE = 2e-3
13
  FACTOR = 22 * 66
@@ -99,7 +99,7 @@ def create_model(tokenizer):
99
  return model
100
 
101
  def load_model():
102
- model = LlamaForCausalLM.from_pretrained(OUTPUT_REPO)
103
  return model
104
 
105
  def configure_tokenizer(tokenizer):
 
3
  import torch
4
  import trl
5
 
6
+ from transformers import AutoTokenizer, LlamaConfig, AutoModelForCasualLM, LlamaForCausalLM, TrainingArguments, PreTrainedTokenizerFast, AdamW, get_cosine_schedule_with_warmup
7
  from datasets import load_dataset, Dataset
8
  from tokenizers import ByteLevelBPETokenizer
9
 
10
+ BATCH_SIZE = 4
11
  EPOCHS = 2
12
  LEARNING_RATE = 2e-3
13
  FACTOR = 22 * 66
 
99
  return model
100
 
101
  def load_model():
102
+ model = AutoModelForCasualLM.from_pretrained(OUTPUT_REPO)
103
  return model
104
 
105
  def configure_tokenizer(tokenizer):