Spaces:
Starting
on
L40S
Starting
on
L40S
Update train.py
Browse files
train.py
CHANGED
|
@@ -3,11 +3,11 @@ import os
|
|
| 3 |
import torch
|
| 4 |
import trl
|
| 5 |
|
| 6 |
-
from transformers import AutoTokenizer, LlamaConfig, LlamaForCausalLM, TrainingArguments, PreTrainedTokenizerFast, AdamW, get_cosine_schedule_with_warmup
|
| 7 |
from datasets import load_dataset, Dataset
|
| 8 |
from tokenizers import ByteLevelBPETokenizer
|
| 9 |
|
| 10 |
-
BATCH_SIZE =
|
| 11 |
EPOCHS = 2
|
| 12 |
LEARNING_RATE = 2e-3
|
| 13 |
FACTOR = 22 * 66
|
|
@@ -99,7 +99,7 @@ def create_model(tokenizer):
|
|
| 99 |
return model
|
| 100 |
|
| 101 |
def load_model():
|
| 102 |
-
model =
|
| 103 |
return model
|
| 104 |
|
| 105 |
def configure_tokenizer(tokenizer):
|
|
|
|
| 3 |
import torch
|
| 4 |
import trl
|
| 5 |
|
| 6 |
+
from transformers import AutoTokenizer, LlamaConfig, AutoModelForCasualLM, LlamaForCausalLM, TrainingArguments, PreTrainedTokenizerFast, AdamW, get_cosine_schedule_with_warmup
|
| 7 |
from datasets import load_dataset, Dataset
|
| 8 |
from tokenizers import ByteLevelBPETokenizer
|
| 9 |
|
| 10 |
+
BATCH_SIZE = 4
|
| 11 |
EPOCHS = 2
|
| 12 |
LEARNING_RATE = 2e-3
|
| 13 |
FACTOR = 22 * 66
|
|
|
|
| 99 |
return model
|
| 100 |
|
| 101 |
def load_model():
|
| 102 |
+
model = AutoModelForCasualLM.from_pretrained(OUTPUT_REPO)
|
| 103 |
return model
|
| 104 |
|
| 105 |
def configure_tokenizer(tokenizer):
|