|
from transformers import GPT2Tokenizer, GPT2LMHeadModel, Trainer, TrainingArguments, DataCollatorForLanguageModeling
|
|
from datasets import Dataset
|
|
|
|
|
|
def load_dataset(file_path):
|
|
with open(file_path, "r", encoding="utf-8") as f:
|
|
text = f.read()
|
|
return [text]
|
|
|
|
|
|
|
|
tokenizer = GPT2Tokenizer.from_pretrained("gpt2")
|
|
tokenizer.pad_token = tokenizer.eos_token
|
|
model = GPT2LMHeadModel.from_pretrained("gpt2")
|
|
|
|
|
|
|
|
model.save_pretrained("./finetuned_gpt2")
|
|
tokenizer.save_pretrained("./finetuned_gpt2")
|
|
|
|
print("Fine-tuning completed.")
|
|
|