Michela
Upload data and app
e62e0c5
"""
This script parses through the cleaned texts and prompts Ollama to create lists of keywords of the contents from each page while focusing on particular topics of interest.
It was tested with German texts from the 17th, 18th, and 19th century.
Code by Michela Vignoli. Parts of this code were developed with assistance from GPT-4 and GPT-3 (free version).
"""
import json
import subprocess
import os
import time
def get_data(root_folder, extension='.txt'):
data = []
# Walk through all folders and files in the root directory
for folder, _, files in os.walk(root_folder):
for file in files:
if file.endswith(extension):
file_path = os.path.join(folder, file)
filename = os.path.splitext(os.path.basename(file_path))[0]
folder_path = os.path.join(f'preprocessed/{folder}_keywords')
#print(f"Found file: {file_path}")
# Read the file content
with open(file_path, 'r', encoding="utf-8") as f:
text = f.read()
data.append({"path": folder_path, "text": text, "filename": filename})
return data
def extract_corrected_text(raw_response):
# Split the response by newline to handle multiple JSON objects
lines = raw_response.strip().split('\n')
response_segments = []
# Process each line as a JSON object
for line in lines:
try:
json_obj = json.loads(line)
response_segment = json_obj.get('response', '')
response_segments.append(response_segment)
except json.JSONDecodeError:
print(f"Skipping invalid JSON line: {line}")
# Join all response segments into a single text
full_text = ''.join(response_segments)
return full_text
def correct_text_with_llm(text, retries=3):
print('-------------------------')
print('Processing LLM request...')
prompt = "You are a historian expert in historical German texts. From the following faulty OCR texts generated from historical traveolgues printed from the 17-19th century, create a list of keywords summarizing the animals, plants, landscapes, and maps mentioned in the text. If animals, plants, landscapes, or maps are mentioned in the text, include the aforementioned general keywords and more specific ones afterwards. If none of these is mentioned in the text, add up to three keywords describing the content of the page.\n\n" + text
# Prepare the cURL command
curl_command = [
'curl',
'http://your.ip:port/api/generate',
'-d', json.dumps({
"model": "llama3.1:70b",
"prompt": prompt
}),
'-H', 'Content-Type: application/json'
]
for attempt in range(retries):
print(f"Attempt {attempt + 1} of {retries}: Processing LLM request...")
result = subprocess.run(curl_command, capture_output=True, text=True, encoding='utf-8')
if result.returncode != 0:
print(f"Error: cURL command failed with error: {result.stderr}")
time.sleep(2) # Wait before retrying
continue
corrected_text = extract_corrected_text(result.stdout)
print("Corrected text:", corrected_text)
if corrected_text:
return corrected_text
print("Invalid response, retrying...")
time.sleep(2) # Wait before retrying
print(f"Failed to get a valid response from the LLM API.")
return None
def process_txt(root_folder):
# Read text files from the root folder
data = get_data(root_folder)
# Process each file in the data list
for item in data:
text = item["text"]
folder_path = item["path"]
filename = item["filename"]
# Define the output path for the corrected text file
output_path = os.path.join(folder_path, f"{filename}_keywords.txt")
# Create the necessary directories if they don't exist
os.makedirs(folder_path, exist_ok=True)
# Apply the LLM correction function
corrected_text = correct_text_with_llm(text)
# Determine the appropriate output path based on whether correction succeeded
if corrected_text:
# Save the corrected text into a .txt file
with open(output_path, "w", encoding="utf-8") as file:
file.write(corrected_text)
print(f"Processed and corrected text saved to {output_path}")
else:
# Save the original text into a .txt file with a different name if correction failed
output_path = os.path.join(folder_path, f"{filename}_FAILED.txt")
with open(output_path, "w", encoding="utf-8") as file:
file.write(text)
print(f"Failed to correct text. Original text saved to {output_path}")
# Example usage
root_folder = 'source/path/'
process_txt(root_folder)