|
import torch
|
|
import torch.nn as nn
|
|
import torch.optim as optim
|
|
|
|
|
|
|
|
def prepare_dataset(filepath, seq_length):
|
|
"""
|
|
Prepares the dataset for training from a text file.
|
|
|
|
Args:
|
|
filepath (str): Path to the text file (e.g., 'dataset.txt').
|
|
seq_length (int): The length of input sequences.
|
|
|
|
Returns:
|
|
tuple: vocab (set), char_to_index (dict), index_to_char (dict),
|
|
input_sequences (list), target_sequences (list)
|
|
"""
|
|
try:
|
|
with open(filepath, 'r', encoding='utf-8') as file:
|
|
text = file.read()
|
|
except FileNotFoundError:
|
|
print(f"Error: File '{filepath}' not found. Make sure the file exists in the correct directory.")
|
|
return None, None, None, None, None
|
|
|
|
vocab = sorted(list(set(text)))
|
|
char_to_index = {char: index for index, char in enumerate(vocab)}
|
|
index_to_char = {index: char for index, char in enumerate(vocab)}
|
|
|
|
input_sequences = []
|
|
target_sequences = []
|
|
|
|
for i in range(0, len(text) - seq_length):
|
|
input_seq = text[i:i + seq_length]
|
|
target_seq = text[i + seq_length]
|
|
input_sequences.append([char_to_index[char] for char in input_seq])
|
|
target_sequences.append(char_to_index[target_seq])
|
|
|
|
return vocab, char_to_index, index_to_char, input_sequences, target_sequences
|
|
|
|
|
|
|
|
|
|
class SimpleRNNLM(nn.Module):
|
|
def __init__(self, vocab_size, embedding_dim, hidden_dim, num_layers):
|
|
super(SimpleRNNLM, self).__init__()
|
|
self.embedding = nn.Embedding(vocab_size, embedding_dim)
|
|
self.rnn = nn.RNN(embedding_dim, hidden_dim, num_layers, batch_first=True)
|
|
self.fc = nn.Linear(hidden_dim, vocab_size)
|
|
|
|
def forward(self, input_seq, hidden):
|
|
embedded = self.embedding(input_seq)
|
|
output, hidden = self.rnn(embedded, hidden)
|
|
output = self.fc(output[:, -1, :])
|
|
return output, hidden
|
|
|
|
def init_hidden(self, batch_size, num_layers, hidden_dim):
|
|
return torch.zeros(num_layers, batch_size, hidden_dim)
|
|
|
|
|
|
|
|
dataset_filepath = 'dataset.txt'
|
|
seq_length = 64
|
|
|
|
vocab, char_to_index, index_to_char, input_seqs, target_seqs = prepare_dataset(dataset_filepath, seq_length)
|
|
|
|
if vocab is None:
|
|
exit()
|
|
|
|
print(f"Vocabulary Size: {len(vocab)}")
|
|
print(f"Number of Input Sequences: {len(input_seqs)}")
|
|
|
|
|
|
|
|
|
|
vocab_size = len(vocab)
|
|
embedding_dim = 32
|
|
hidden_dim = 64
|
|
num_layers = 1
|
|
learning_rate = 0.01
|
|
num_epochs = 10
|
|
|
|
model = SimpleRNNLM(vocab_size, embedding_dim, hidden_dim, num_layers)
|
|
criterion = nn.CrossEntropyLoss()
|
|
optimizer = optim.Adam(model.parameters(), lr=learning_rate)
|
|
|
|
device = torch.device("cpu")
|
|
model.to(device)
|
|
criterion.to(device)
|
|
|
|
|
|
|
|
|
|
batch_size = 256
|
|
|
|
for epoch in range(num_epochs):
|
|
model.train()
|
|
total_loss = 0
|
|
|
|
for i in range(0, len(input_seqs), batch_size):
|
|
input_batch = input_seqs[i:i+batch_size]
|
|
target_batch = target_seqs[i:i+batch_size]
|
|
|
|
input_batch_tensor = torch.LongTensor(input_batch).to(device)
|
|
target_batch_tensor = torch.LongTensor(target_batch).to(device)
|
|
|
|
hidden = model.init_hidden(len(input_batch), num_layers, hidden_dim).to(device)
|
|
|
|
optimizer.zero_grad()
|
|
|
|
output, hidden = model(input_batch_tensor, hidden)
|
|
loss = criterion(output, target_batch_tensor)
|
|
|
|
loss.backward()
|
|
optimizer.step()
|
|
|
|
total_loss += loss.item()
|
|
|
|
average_loss = total_loss / (len(input_seqs) // batch_size + (len(input_seqs) % batch_size != 0))
|
|
print(f"Epoch [{epoch+1}/{num_epochs}], Loss: {average_loss:.4f}")
|
|
|
|
|
|
|
|
|
|
def generate_text(model, start_text, predict_len, char_to_index, index_to_char, vocab, device):
|
|
model.eval()
|
|
generated_text = start_text
|
|
|
|
input_sequence = [char_to_index[char] for char in start_text]
|
|
input_tensor = torch.LongTensor([input_sequence]).to(device)
|
|
|
|
hidden = model.init_hidden(1, num_layers, hidden_dim).to(device)
|
|
|
|
with torch.no_grad():
|
|
for _ in range(predict_len):
|
|
output, hidden = model(input_tensor, hidden)
|
|
|
|
probabilities = torch.softmax(output, dim=1)
|
|
predicted_index = torch.multinomial(probabilities, 1).item()
|
|
predicted_char = index_to_char[predicted_index]
|
|
|
|
generated_text += predicted_char
|
|
|
|
input_sequence = input_sequence[1:] + [predicted_index]
|
|
input_tensor = torch.LongTensor([input_sequence]).to(device)
|
|
|
|
return generated_text
|
|
|
|
|
|
|
|
start_text = "The "
|
|
predict_length = 500
|
|
|
|
generated_output = generate_text(model, start_text, predict_length, char_to_index, index_to_char, vocab, device)
|
|
print("\nGenerated Text:")
|
|
print(generated_output) |