Spaces:
Runtime error
Runtime error
Update train.py
Browse files
train.py
CHANGED
@@ -3,11 +3,11 @@ import os
|
|
3 |
import torch
|
4 |
import trl
|
5 |
|
6 |
-
from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM, TrainingArguments, PreTrainedTokenizerFast, AdamW, get_cosine_schedule_with_warmup
|
7 |
from datasets import load_dataset, Dataset
|
8 |
from tokenizers import ByteLevelBPETokenizer
|
9 |
|
10 |
-
BATCH_SIZE =
|
11 |
EPOCHS = 2
|
12 |
LEARNING_RATE = 2e-3
|
13 |
FACTOR = 22 * 66
|
@@ -99,7 +99,7 @@ def create_model(tokenizer):
|
|
99 |
return model
|
100 |
|
101 |
def load_model():
|
102 |
-
model =
|
103 |
return model
|
104 |
|
105 |
def configure_tokenizer(tokenizer):
|
|
|
3 |
import torch
|
4 |
import trl
|
5 |
|
6 |
+
from transformers import AutoTokenizer, LlamaConfig, AutoModelForCasualLM, LlamaForCausalLM, TrainingArguments, PreTrainedTokenizerFast, AdamW, get_cosine_schedule_with_warmup
|
7 |
from datasets import load_dataset, Dataset
|
8 |
from tokenizers import ByteLevelBPETokenizer
|
9 |
|
10 |
+
BATCH_SIZE = 4
|
11 |
EPOCHS = 2
|
12 |
LEARNING_RATE = 2e-3
|
13 |
FACTOR = 22 * 66
|
|
|
99 |
return model
|
100 |
|
101 |
def load_model():
|
102 |
+
model = AutoModelForCasualLM.from_pretrained(OUTPUT_REPO)
|
103 |
return model
|
104 |
|
105 |
def configure_tokenizer(tokenizer):
|