File size: 13,130 Bytes
9be760a
 
 
84b09e2
a8a8bc2
6c0aa26
96b58ba
 
9be760a
 
7445a27
e8bf2aa
96b58ba
 
 
 
396d07b
96b58ba
9be760a
6c0aa26
c1f45eb
6c0aa26
96b58ba
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d07d4c9
 
96b58ba
 
 
 
 
 
 
c1f45eb
96b58ba
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9983d3c
96b58ba
 
 
 
 
 
 
 
 
 
 
 
 
9983d3c
96b58ba
 
 
 
 
 
 
 
 
9be760a
 
96b58ba
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
import gradio as gr
import json
import re
import sqlite3
import logging
from collections import defaultdict
from typing import Tuple, Dict, List

from util import process_json_files
from gematria import calculate_gematria
from deep_translator import GoogleTranslator, exceptions
from urllib.parse import quote_plus
from tqdm import tqdm # Import tqdm for progress bars

# Constants
DATABASE_FILE = 'gematria.db'
MAX_PHRASE_LENGTH_LIMIT = 20 # Populate database for phrases up to 5 words
BATCH_SIZE = 1000 # Insert phrases into database in batches

# Set up logging
logging.basicConfig(level=logging.DEBUG, format='%(asctime)s - %(levelname)s - %(filename)s - %(lineno)d - %(message)s')

# Global variables
conn: sqlite3.Connection = None
translator: GoogleTranslator = None
book_names: Dict[int, str] = {}
gematria_cache: Dict[Tuple[int, int], List[Tuple[str, str, int, int]]] = {}
translation_cache: Dict[str, str] = {}

def initialize_database() -> None:
  """Initializes the SQLite database."""
  global conn
  conn = sqlite3.connect(DATABASE_FILE, isolation_level=None) # Autocommit for faster insertion
  cursor = conn.cursor()

  # Create tables if they don't exist
  cursor.execute('''
  CREATE TABLE IF NOT EXISTS results (
    gematria_sum INTEGER,
    words TEXT,
    translation TEXT,
    book TEXT,
    chapter INTEGER,
    verse INTEGER,
    PRIMARY KEY (gematria_sum, words, book, chapter, verse)
  )
  ''')
  cursor.execute('''
  CREATE TABLE IF NOT EXISTS processed_books (
    book TEXT PRIMARY KEY,
    max_phrase_length INTEGER
  )
  ''')
  cursor.execute('''
  CREATE TABLE IF NOT EXISTS translations (
    hebrew_phrase TEXT PRIMARY KEY,
    english_translation TEXT
  )
  ''')

def initialize_translator() -> None:
  """Initializes the Google Translator."""
  global translator
  translator = GoogleTranslator(source='iw', target='en')
  logging.info("Translator initialized.")

def populate_database(start_book: int, end_book: int, max_phrase_length: int = 1) -> None:
  """Populates the database with phrases from the Tanach and their Gematria values."""
  global conn, book_names
  logging.info(f"Populating database with books from {start_book} to {end_book}...")
  cursor = conn.cursor()

  for book_id in tqdm(range(start_book, end_book + 1), desc="Processing Books"):
    book_data = process_json_files(book_id, book_id) # Get data for the single book

    # process_json_files returns a dictionary with book_id as key,
    # so access the book data directly
    if book_id in book_data:
      book_data = book_data[book_id]
      if 'title' not in book_data or not isinstance(book_data['title'], str):
        logging.warning(f"Skipping book {book_id} due to missing or invalid 'title' field.")
        continue

      title = book_data['title']
      book_names[book_id] = title

      # Check if the book is already processed for this max_phrase_length
      cursor.execute('''SELECT max_phrase_length FROM processed_books WHERE book = ?''', (title,))
      result = cursor.fetchone()
      if result and result[0] >= max_phrase_length:
        logging.info(f"Skipping book {title}: Already processed with max_phrase_length {result[0]}")
        continue

      logging.info(f"Processing book {title} with max_phrase_length {max_phrase_length}")

      if 'text' not in book_data or not isinstance(book_data['text'], list):
        logging.warning(f"Skipping book {book_id} due to missing or invalid 'text' field.")
        continue

      chapters = book_data['text']
      # Faster iteration with enumerate and list comprehension
      for chapter_id, chapter in enumerate(chapters):
        for verse_id, verse in enumerate(chapter):
          verse_text = flatten_text(verse)
          # Remove text in square brackets and non-Hebrew characters
          verse_text = re.sub(r'\[.*?\]', '', verse_text)
          verse_text = re.sub(r"[^\u05D0-\u05EA ]+", "", verse_text)
          verse_text = re.sub(r" +", " ", verse_text)
          words = verse_text.split()

          # Use a generator to avoid building large lists in memory
          for length in range(1, max_phrase_length + 1):
            for start in range(len(words) - length + 1):
              phrase_candidate = " ".join(words[start:start + length])
              gematria_sum = calculate_gematria(phrase_candidate.replace(" ", ""))
              yield gematria_sum, phrase_candidate, title, chapter_id + 1, verse_id + 1

      # Mark the book as processed with the current max_phrase_length
      cursor.execute('''
      INSERT OR REPLACE INTO processed_books (book, max_phrase_length)
      VALUES (?, ?)
      ''', (title, max_phrase_length))

def insert_phrases_to_db(phrases: List[Tuple[int, str, str, int, int]]) -> None:
  """Inserts a list of phrases into the database efficiently."""
  global conn
  cursor = conn.cursor()

  # Use executemany to insert multiple rows at once
  cursor.executemany('''
  INSERT OR IGNORE INTO results (gematria_sum, words, book, chapter, verse)
  VALUES (?, ?, ?, ?, ?)
  ''', phrases)

  # Commit the changes outside the loop for better performance
  conn.commit()

def get_translation(phrase: str) -> str:
  """Retrieves or generates the English translation of a Hebrew phrase."""
  global translator, conn, translation_cache
  if phrase in translation_cache:
    return translation_cache[phrase]
  else:
    cursor = conn.cursor()
    cursor.execute('''
    SELECT english_translation FROM translations
    WHERE hebrew_phrase = ?
    ''', (phrase,))
    result = cursor.fetchone()
    if result and result[0]:
      translation = result[0]
      return translation
    else:
      translation = translate_and_store(phrase)
      cursor.execute('''
      INSERT OR IGNORE INTO translations (hebrew_phrase, english_translation)
      VALUES (?, ?)
      ''', (phrase, translation))
      return translation

def translate_and_store(phrase: str) -> str:
  """Translates a Hebrew phrase to English using Google Translate and handles potential errors."""
  global translator
  max_retries = 3
  retries = 0

  while retries < max_retries:
    try:
      translation = translator.translate(phrase)
      logging.debug(f"Translated phrase: {translation}")
      return translation
    except (exceptions.TranslationNotFound, exceptions.NotValidPayload,
         exceptions.ServerException, exceptions.RequestError, requests.exceptions.ConnectionError) as e:
      retries += 1
      logging.warning(f"Error translating phrase '{phrase}': {e}. Retrying... ({retries}/{max_retries})")

  logging.error(f"Failed to translate phrase '{phrase}' after {max_retries} retries.")
  return "[Translation Error]"

def search_gematria_in_db(gematria_sum: int, max_words: int) -> List[Tuple[str, str, int, int]]:
  """Searches the database for phrases with a given Gematria value and word count.
    Returns phrases with word count <= max_words."""
  global conn
  cursor = conn.cursor()
  logging.debug(f"Searching for phrases with Gematria: {gematria_sum} and max words: {max_words}")
  cursor.execute('''
  SELECT words, book, chapter, verse FROM results WHERE gematria_sum = ?
  ''', (gematria_sum,)) # Retrieve all matching phrases first
  results = cursor.fetchall()
  filtered_results = []
  logging.debug(f"Found {len(results)} matching phrases before filtering.")
  for words, book, chapter, verse in results:
    # Filter by word count (including phrases with fewer words)
    word_count = len(words.split()) # Correctly split and count words
    logging.debug(f"Word count for '{words}': {word_count}")
    if word_count <= max_words: # Include phrases with word count <= max_words
      filtered_results.append((words, book, chapter, verse))
  logging.debug(f"Found {len(filtered_results)} matching phrases after filtering.")
  return filtered_results

def gematria_search_interface(phrase: str, max_words: int, show_translation: bool) -> str:
  """The main function for the Gradio interface."""
  if not phrase.strip():
    return "Please enter a phrase."

  global conn, book_names, gematria_cache
  conn = sqlite3.connect(DATABASE_FILE)
  cursor = conn.cursor()

  # Extract numbers from the input text
  numbers = re.findall(r'\d+', phrase)
  # Calculate Gematria for the remaining text (non-numbers)
  text_without_numbers = re.sub(r'\d+', '', phrase)
  phrase_gematria = calculate_gematria(text_without_numbers.replace(" ", ""))

  # Add sum of numbers to Gematria
  phrase_gematria += sum(int(number) for number in numbers)

  logging.info(f"Searching for phrases with Gematria: {phrase_gematria}")

  # Debugging output
  logging.debug(f"Phrase Gematria: {phrase_gematria}")
  logging.debug(f"Max Words: {max_words}")

  # Check if Gematria is in cache for the specific max_words value
  if (phrase_gematria, max_words) in gematria_cache:
    matching_phrases = gematria_cache[(phrase_gematria, max_words)]
    logging.debug(f"Retrieved matching phrases from cache for max_words: {max_words}.")
  else:
    # Search in the database
    matching_phrases = search_gematria_in_db(phrase_gematria, max_words)
    # Cache the results with the max_words value
    gematria_cache[(phrase_gematria, max_words)] = matching_phrases
    logging.debug(f"Retrieved matching phrases from database for max_words: {max_words}.")

  if not matching_phrases:
    return "No matching phrases found."

  # Sort results by book, chapter, and verse
  sorted_phrases = sorted(matching_phrases, key=lambda x: (int(list(book_names.keys())[list(book_names.values()).index(x[1])]), x[2], x[3]))
  logging.debug(f"Sorted matching phrases: {sorted_phrases}")

  # Group results by book
  results_by_book = defaultdict(list)
  for words, book, chapter, verse in sorted_phrases:
    results_by_book[book].append((words, chapter, verse))
  logging.debug(f"Grouped results by book: {results_by_book}")

  # Format results for display
  results = []
  results.append("<div class='results-container'>")
  for book, phrases in results_by_book.items():
    results.append(f"<h4>Book: {book}</h4>") # Directly display book name
    for words, chapter, verse in phrases:
      translation = get_translation(words) if show_translation else ""
      link = f"https://www.biblegateway.com/passage/?search={quote_plus(book)}+{chapter}%3A{verse}&version=CJB"
      results.append(f"""
      <div class='result-item'>
        <p>Chapter: {chapter}, Verse: {verse}</p>
        <p class='hebrew-phrase'>Hebrew Phrase: {words}</p>
        <p>Translation: {translation}</p>
        <a href='{link}' target='_blank' class='bible-link'>[See on Bible Gateway]</a>
      </div>
      """)
  results.append("</div>") # Close results-container div

  conn.close()

  # Add CSS styling
  style = """
  <style>
    .results-container {
      display: grid;
      grid-template-columns: repeat(auto-fit, minmax(300px, 1fr));
      gap: 20px;
    }

    .result-item {
      border: 1px solid #ccc;
      padding: 15px;
      border-radius: 5px;
      box-shadow: 2px 2px 5px rgba(0, 0, 0, 0.1);
    }

    .hebrew-phrase {
      font-family: 'SBL Hebrew', 'Ezra SIL', serif;
      direction: rtl;
    }

    .bible-link {
      display: block;
      margin-top: 10px;
      color: #007bff;
      text-decoration: none;
    }
  </style>
  """

  return style + "\n".join(results)

def flatten_text(text: List) -> str:
  """Helper function to flatten nested lists into a single list."""
  if isinstance(text, list):
    return " ".join(flatten_text(item) if isinstance(item, list) else item for item in text)
  return text

def run_app() -> None:
  """Initializes and launches the Gradio app."""
  initialize_database()
  initialize_translator()

  # Pre-populate the database
  logging.info("Starting database population...")
  phrases_to_insert = [] # Collect phrases before inserting in bulk
  for max_phrase_length in range(1, MAX_PHRASE_LENGTH_LIMIT + 1): # Populate for phrases up to MAX_PHRASE_LENGTH_LIMIT words
    for gematria_sum, phrase, book, chapter, verse in tqdm(populate_database(1, 39, max_phrase_length=max_phrase_length), desc=f"Populating Database (Max Length: {max_phrase_length})"): # Books 1 to 39
      phrases_to_insert.append((gematria_sum, phrase, book, chapter, verse))
      if len(phrases_to_insert) >= BATCH_SIZE: # Insert in batches of BATCH_SIZE for efficiency
        insert_phrases_to_db(phrases_to_insert)
        phrases_to_insert = []
    if phrases_to_insert: # Insert remaining phrases
      insert_phrases_to_db(phrases_to_insert)
  logging.info("Database population complete.")

  iface = gr.Interface(
    fn=gematria_search_interface,
    inputs=[
      gr.Textbox(label="Enter word(s) or numbers (e.g., 'abc', '888' or 'abc 111 777')"),
      gr.Number(label="Max Word Count in Result Phrases", value=1, minimum=1, maximum=MAX_PHRASE_LENGTH_LIMIT),
      gr.Checkbox(label="Show Translation", value=True)
    ],
    outputs=gr.HTML(label="Results"),
    title="Gematria Search in Tanach",
    description="Search for phrases and/or numbers in the Tanach that have the same Gematria value.",
    live=False,
    allow_flagging="never"
  )
  iface.launch()

if __name__ == "__main__":
  run_app()