code
stringlengths
66
870k
docstring
stringlengths
19
26.7k
func_name
stringlengths
1
138
language
stringclasses
1 value
repo
stringlengths
7
68
path
stringlengths
5
324
url
stringlengths
46
389
license
stringclasses
7 values
def cleanup_error_artifacts(uuid, datastore): """Helper function to clean up error artifacts""" cleanup_files = ["last-error-screenshot.png", "last-error.txt"] for f in cleanup_files: full_path = os.path.join(datastore.datastore_path, uuid, f) if os.path.isfile(full_path): os.unlink(full_path)
Helper function to clean up error artifacts
cleanup_error_artifacts
python
dgtlmoon/changedetection.io
changedetectionio/async_update_worker.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/async_update_worker.py
Apache-2.0
async def send_content_changed_notification(watch_uuid, notification_q, datastore): """Helper function to queue notifications using the new notification service""" try: from changedetectionio.notification_service import create_notification_service # Create notification service instance notification_service = create_notification_service(datastore, notification_q) notification_service.send_content_changed_notification(watch_uuid) except Exception as e: logger.error(f"Error sending notification for {watch_uuid}: {e}")
Helper function to queue notifications using the new notification service
send_content_changed_notification
python
dgtlmoon/changedetection.io
changedetectionio/async_update_worker.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/async_update_worker.py
Apache-2.0
async def send_filter_failure_notification(watch_uuid, notification_q, datastore): """Helper function to send filter failure notifications using the new notification service""" try: from changedetectionio.notification_service import create_notification_service # Create notification service instance notification_service = create_notification_service(datastore, notification_q) notification_service.send_filter_failure_notification(watch_uuid) except Exception as e: logger.error(f"Error sending filter failure notification for {watch_uuid}: {e}")
Helper function to send filter failure notifications using the new notification service
send_filter_failure_notification
python
dgtlmoon/changedetection.io
changedetectionio/async_update_worker.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/async_update_worker.py
Apache-2.0
async def send_step_failure_notification(watch_uuid, step_n, notification_q, datastore): """Helper function to send step failure notifications using the new notification service""" try: from changedetectionio.notification_service import create_notification_service # Create notification service instance notification_service = create_notification_service(datastore, notification_q) notification_service.send_step_failure_notification(watch_uuid, step_n) except Exception as e: logger.error(f"Error sending step failure notification for {watch_uuid}: {e}")
Helper function to send step failure notifications using the new notification service
send_step_failure_notification
python
dgtlmoon/changedetection.io
changedetectionio/async_update_worker.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/async_update_worker.py
Apache-2.0
def login_optionally_required(func): """ If password authentication is enabled, verify the user is logged in. To be used as a decorator for routes that should optionally require login. This version is blueprint-friendly as it uses current_app instead of directly accessing app. """ @wraps(func) def decorated_view(*args, **kwargs): from flask import current_app import flask_login from flask_login import current_user # Access datastore through the app config datastore = current_app.config['DATASTORE'] has_password_enabled = datastore.data['settings']['application'].get('password') or os.getenv("SALTED_PASS", False) # Permitted if request.endpoint and 'diff_history_page' in request.endpoint and datastore.data['settings']['application'].get('shared_diff_access'): return func(*args, **kwargs) elif request.method in flask_login.config.EXEMPT_METHODS: return func(*args, **kwargs) elif current_app.config.get('LOGIN_DISABLED'): return func(*args, **kwargs) elif has_password_enabled and not current_user.is_authenticated: return current_app.login_manager.unauthorized() return func(*args, **kwargs) return decorated_view
If password authentication is enabled, verify the user is logged in. To be used as a decorator for routes that should optionally require login. This version is blueprint-friendly as it uses current_app instead of directly accessing app.
login_optionally_required
python
dgtlmoon/changedetection.io
changedetectionio/auth_decorator.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/auth_decorator.py
Apache-2.0
def get_uuid_position(self, target_uuid): """ Find the position of a watch UUID in the priority queue. Optimized for large queues - O(n) complexity instead of O(n log n). Args: target_uuid: The UUID to search for Returns: dict: Contains position info or None if not found - position: 0-based position in queue (0 = next to be processed) - total_items: total number of items in queue - priority: the priority value of the found item """ with self.mutex: queue_list = list(self.queue) total_items = len(queue_list) if total_items == 0: return { 'position': None, 'total_items': 0, 'priority': None, 'found': False } # Find the target item and its priority first - O(n) target_item = None target_priority = None for item in queue_list: if (hasattr(item, 'item') and isinstance(item.item, dict) and item.item.get('uuid') == target_uuid): target_item = item target_priority = item.priority break if target_item is None: return { 'position': None, 'total_items': total_items, 'priority': None, 'found': False } # Count how many items have higher priority (lower numbers) - O(n) position = 0 for item in queue_list: # Items with lower priority numbers are processed first if item.priority < target_priority: position += 1 elif item.priority == target_priority and item != target_item: # For same priority, count items that come before this one # (Note: this is approximate since heap order isn't guaranteed for equal priorities) position += 1 return { 'position': position, 'total_items': total_items, 'priority': target_priority, 'found': True }
Find the position of a watch UUID in the priority queue. Optimized for large queues - O(n) complexity instead of O(n log n). Args: target_uuid: The UUID to search for Returns: dict: Contains position info or None if not found - position: 0-based position in queue (0 = next to be processed) - total_items: total number of items in queue - priority: the priority value of the found item
get_uuid_position
python
dgtlmoon/changedetection.io
changedetectionio/custom_queue.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/custom_queue.py
Apache-2.0
def get_queue_summary(self): """ Get a quick summary of queue state without expensive operations. O(n) complexity - fast even for large queues. Returns: dict: Queue summary statistics """ with self.mutex: queue_list = list(self.queue) total_items = len(queue_list) if total_items == 0: return { 'total_items': 0, 'priority_breakdown': {}, 'immediate_items': 0, 'clone_items': 0, 'scheduled_items': 0 } # Count items by priority type - O(n) immediate_items = 0 # priority 1 clone_items = 0 # priority 5 scheduled_items = 0 # priority > 100 (timestamps) priority_counts = {} for item in queue_list: priority = item.priority priority_counts[priority] = priority_counts.get(priority, 0) + 1 if priority == 1: immediate_items += 1 elif priority == 5: clone_items += 1 elif priority > 100: scheduled_items += 1 return { 'total_items': total_items, 'priority_breakdown': priority_counts, 'immediate_items': immediate_items, 'clone_items': clone_items, 'scheduled_items': scheduled_items, 'min_priority': min(priority_counts.keys()) if priority_counts else None, 'max_priority': max(priority_counts.keys()) if priority_counts else None }
Get a quick summary of queue state without expensive operations. O(n) complexity - fast even for large queues. Returns: dict: Queue summary statistics
get_queue_summary
python
dgtlmoon/changedetection.io
changedetectionio/custom_queue.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/custom_queue.py
Apache-2.0
def get_uuid_position(self, target_uuid): """ Find the position of a watch UUID in the async priority queue. Optimized for large queues - O(n) complexity instead of O(n log n). Args: target_uuid: The UUID to search for Returns: dict: Contains position info or None if not found - position: 0-based position in queue (0 = next to be processed) - total_items: total number of items in queue - priority: the priority value of the found item """ queue_list = list(self._queue) total_items = len(queue_list) if total_items == 0: return { 'position': None, 'total_items': 0, 'priority': None, 'found': False } # Find the target item and its priority first - O(n) target_item = None target_priority = None for item in queue_list: if (hasattr(item, 'item') and isinstance(item.item, dict) and item.item.get('uuid') == target_uuid): target_item = item target_priority = item.priority break if target_item is None: return { 'position': None, 'total_items': total_items, 'priority': None, 'found': False } # Count how many items have higher priority (lower numbers) - O(n) position = 0 for item in queue_list: if item.priority < target_priority: position += 1 elif item.priority == target_priority and item != target_item: position += 1 return { 'position': position, 'total_items': total_items, 'priority': target_priority, 'found': True }
Find the position of a watch UUID in the async priority queue. Optimized for large queues - O(n) complexity instead of O(n log n). Args: target_uuid: The UUID to search for Returns: dict: Contains position info or None if not found - position: 0-based position in queue (0 = next to be processed) - total_items: total number of items in queue - priority: the priority value of the found item
get_uuid_position
python
dgtlmoon/changedetection.io
changedetectionio/custom_queue.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/custom_queue.py
Apache-2.0
def get_queue_summary(self): """ Get a quick summary of async queue state. O(n) complexity - fast even for large queues. """ queue_list = list(self._queue) total_items = len(queue_list) if total_items == 0: return { 'total_items': 0, 'priority_breakdown': {}, 'immediate_items': 0, 'clone_items': 0, 'scheduled_items': 0 } immediate_items = 0 clone_items = 0 scheduled_items = 0 priority_counts = {} for item in queue_list: priority = item.priority priority_counts[priority] = priority_counts.get(priority, 0) + 1 if priority == 1: immediate_items += 1 elif priority == 5: clone_items += 1 elif priority > 100: scheduled_items += 1 return { 'total_items': total_items, 'priority_breakdown': priority_counts, 'immediate_items': immediate_items, 'clone_items': clone_items, 'scheduled_items': scheduled_items, 'min_priority': min(priority_counts.keys()) if priority_counts else None, 'max_priority': max(priority_counts.keys()) if priority_counts else None }
Get a quick summary of async queue state. O(n) complexity - fast even for large queues.
get_queue_summary
python
dgtlmoon/changedetection.io
changedetectionio/custom_queue.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/custom_queue.py
Apache-2.0
def customSequenceMatcher( before: List[str], after: List[str], include_equal: bool = False, include_removed: bool = True, include_added: bool = True, include_replaced: bool = True, include_change_type_prefix: bool = True, html_colour: bool = False ) -> Iterator[List[str]]: """ Compare two sequences and yield differences based on specified parameters. Args: before (List[str]): Original sequence after (List[str]): Modified sequence include_equal (bool): Include unchanged parts include_removed (bool): Include removed parts include_added (bool): Include added parts include_replaced (bool): Include replaced parts include_change_type_prefix (bool): Add prefixes to indicate change types html_colour (bool): Use HTML background colors for differences Yields: List[str]: Differences between sequences """ cruncher = difflib.SequenceMatcher(isjunk=lambda x: x in " \t", a=before, b=after) for tag, alo, ahi, blo, bhi in cruncher.get_opcodes(): if include_equal and tag == 'equal': yield before[alo:ahi] elif include_removed and tag == 'delete': if html_colour: yield [f'<span style="{REMOVED_STYLE}">{line}</span>' for line in same_slicer(before, alo, ahi)] else: yield [f"(removed) {line}" for line in same_slicer(before, alo, ahi)] if include_change_type_prefix else same_slicer(before, alo, ahi) elif include_replaced and tag == 'replace': if html_colour: yield [f'<span style="{REMOVED_STYLE}">{line}</span>' for line in same_slicer(before, alo, ahi)] + \ [f'<span style="{ADDED_STYLE}">{line}</span>' for line in same_slicer(after, blo, bhi)] else: yield [f"(changed) {line}" for line in same_slicer(before, alo, ahi)] + \ [f"(into) {line}" for line in same_slicer(after, blo, bhi)] if include_change_type_prefix else same_slicer(before, alo, ahi) + same_slicer(after, blo, bhi) elif include_added and tag == 'insert': if html_colour: yield [f'<span style="{ADDED_STYLE}">{line}</span>' for line in same_slicer(after, blo, bhi)] else: yield [f"(added) {line}" for line in same_slicer(after, blo, bhi)] if include_change_type_prefix else same_slicer(after, blo, bhi)
Compare two sequences and yield differences based on specified parameters. Args: before (List[str]): Original sequence after (List[str]): Modified sequence include_equal (bool): Include unchanged parts include_removed (bool): Include removed parts include_added (bool): Include added parts include_replaced (bool): Include replaced parts include_change_type_prefix (bool): Add prefixes to indicate change types html_colour (bool): Use HTML background colors for differences Yields: List[str]: Differences between sequences
customSequenceMatcher
python
dgtlmoon/changedetection.io
changedetectionio/diff.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/diff.py
Apache-2.0
def render_diff( previous_version_file_contents: str, newest_version_file_contents: str, include_equal: bool = False, include_removed: bool = True, include_added: bool = True, include_replaced: bool = True, line_feed_sep: str = "\n", include_change_type_prefix: bool = True, patch_format: bool = False, html_colour: bool = False ) -> str: """ Render the difference between two file contents. Args: previous_version_file_contents (str): Original file contents newest_version_file_contents (str): Modified file contents include_equal (bool): Include unchanged parts include_removed (bool): Include removed parts include_added (bool): Include added parts include_replaced (bool): Include replaced parts line_feed_sep (str): Separator for lines in output include_change_type_prefix (bool): Add prefixes to indicate change types patch_format (bool): Use patch format for output html_colour (bool): Use HTML background colors for differences Returns: str: Rendered difference """ newest_lines = [line.rstrip() for line in newest_version_file_contents.splitlines()] previous_lines = [line.rstrip() for line in previous_version_file_contents.splitlines()] if previous_version_file_contents else [] if patch_format: patch = difflib.unified_diff(previous_lines, newest_lines) return line_feed_sep.join(patch) rendered_diff = customSequenceMatcher( before=previous_lines, after=newest_lines, include_equal=include_equal, include_removed=include_removed, include_added=include_added, include_replaced=include_replaced, include_change_type_prefix=include_change_type_prefix, html_colour=html_colour ) def flatten(lst: List[Union[str, List[str]]]) -> str: return line_feed_sep.join(flatten(x) if isinstance(x, list) else x for x in lst) return flatten(rendered_diff)
Render the difference between two file contents. Args: previous_version_file_contents (str): Original file contents newest_version_file_contents (str): Modified file contents include_equal (bool): Include unchanged parts include_removed (bool): Include removed parts include_added (bool): Include added parts include_replaced (bool): Include replaced parts line_feed_sep (str): Separator for lines in output include_change_type_prefix (bool): Add prefixes to indicate change types patch_format (bool): Use patch format for output html_colour (bool): Use HTML background colors for differences Returns: str: Rendered difference
render_diff
python
dgtlmoon/changedetection.io
changedetectionio/diff.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/diff.py
Apache-2.0
def get_socketio_path(): """Generate the correct Socket.IO path prefix for the client""" # If behind a proxy with a sub-path, we need to respect that path prefix = "" if os.getenv('USE_X_SETTINGS') and 'X-Forwarded-Prefix' in request.headers: prefix = request.headers['X-Forwarded-Prefix'] # Socket.IO will be available at {prefix}/socket.io/ return prefix
Generate the correct Socket.IO path prefix for the client
get_socketio_path
python
dgtlmoon/changedetection.io
changedetectionio/flask_app.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/flask_app.py
Apache-2.0
def _jinja2_filter_format_number_locale(value: float) -> str: "Formats for example 4000.10 to the local locale default of 4,000.10" # Format the number with two decimal places (locale format string will return 6 decimal) formatted_value = locale.format_string("%.2f", value, grouping=True) return formatted_value
Formats for example 4000.10 to the local locale default of 4,000.10
_jinja2_filter_format_number_locale
python
dgtlmoon/changedetection.io
changedetectionio/flask_app.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/flask_app.py
Apache-2.0
def _get_worker_status_info(): """Get detailed worker status information for display""" status = worker_handler.get_worker_status() running_uuids = worker_handler.get_running_uuids() return { 'count': status['worker_count'], 'type': status['worker_type'], 'active_workers': len(running_uuids), 'processing_watches': running_uuids, 'loop_running': status.get('async_loop_running', None) }
Get detailed worker status information for display
_get_worker_status_info
python
dgtlmoon/changedetection.io
changedetectionio/flask_app.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/flask_app.py
Apache-2.0
def process_formdata(self, valuelist): """ Processes the raw input from the form and stores it as a string. """ if valuelist: time_str = valuelist[0] # Simple validation for HH:MM format if not time_str or len(time_str.split(":")) != 2: raise ValidationError("Invalid time format. Use HH:MM.") self.data = time_str
Processes the raw input from the form and stores it as a string.
process_formdata
python
dgtlmoon/changedetection.io
changedetectionio/forms.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/forms.py
Apache-2.0
def memory_cleanup(app=None): """ Perform comprehensive memory cleanup operations and log memory usage at each step with nicely formatted numbers. Args: app: Optional Flask app instance for clearing Flask-specific caches Returns: str: Status message """ # Get current process process = psutil.Process() # Log initial memory usage with nicely formatted numbers current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"Memory cleanup started - Current memory usage: {current_memory:,.2f} MB") # 1. Standard garbage collection - force full collection on all generations gc.collect(0) # Collect youngest generation gc.collect(1) # Collect middle generation gc.collect(2) # Collect oldest generation # Run full collection again to ensure maximum cleanup gc.collect() current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After full gc.collect() - Memory usage: {current_memory:,.2f} MB") # 3. Call libc's malloc_trim to release memory back to the OS libc = ctypes.CDLL("libc.so.6") libc.malloc_trim(0) current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After malloc_trim(0) - Memory usage: {current_memory:,.2f} MB") # 4. Clear Python's regex cache re.purge() current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After re.purge() - Memory usage: {current_memory:,.2f} MB") # 5. Reset thread-local storage # Create a new thread local object to encourage cleanup of old ones threading.local() current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After threading.local() - Memory usage: {current_memory:,.2f} MB") # 6. Clear sys.intern cache if Python version supports it try: sys.intern.clear() current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After sys.intern.clear() - Memory usage: {current_memory:,.2f} MB") except (AttributeError, TypeError): logger.debug("sys.intern.clear() not supported in this Python version") # 7. Clear XML/lxml caches if available try: # Check if lxml.etree is in use lxml_etree = sys.modules.get('lxml.etree') if lxml_etree: # Clear module-level caches if hasattr(lxml_etree, 'clear_error_log'): lxml_etree.clear_error_log() # Check for _ErrorLog and _RotatingErrorLog objects and clear them for obj in gc.get_objects(): if hasattr(obj, '__class__') and hasattr(obj.__class__, '__name__'): class_name = obj.__class__.__name__ if class_name in ('_ErrorLog', '_RotatingErrorLog', '_DomainErrorLog') and hasattr(obj, 'clear'): try: obj.clear() except (AttributeError, TypeError): pass # Clear Element objects which can hold references to documents elif class_name in ('_Element', 'ElementBase') and hasattr(obj, 'clear'): try: obj.clear() except (AttributeError, TypeError): pass current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After lxml.etree cleanup - Memory usage: {current_memory:,.2f} MB") # Check if lxml.html is in use lxml_html = sys.modules.get('lxml.html') if lxml_html: # Clear HTML-specific element types for obj in gc.get_objects(): if hasattr(obj, '__class__') and hasattr(obj.__class__, '__name__'): class_name = obj.__class__.__name__ if class_name in ('HtmlElement', 'FormElement', 'InputElement', 'SelectElement', 'TextareaElement', 'CheckboxGroup', 'RadioGroup', 'MultipleSelectOptions', 'FieldsDict') and hasattr(obj, 'clear'): try: obj.clear() except (AttributeError, TypeError): pass current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After lxml.html cleanup - Memory usage: {current_memory:,.2f} MB") except (ImportError, AttributeError): logger.debug("lxml cleanup not applicable") # 8. Clear JSON parser caches if applicable try: # Check if json module is being used and try to clear its cache json_module = sys.modules.get('json') if json_module and hasattr(json_module, '_default_encoder'): json_module._default_encoder.markers.clear() current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After JSON parser cleanup - Memory usage: {current_memory:,.2f} MB") except (AttributeError, KeyError): logger.debug("JSON cleanup not applicable") # 9. Force Python's memory allocator to release unused memory try: if hasattr(sys, 'pypy_version_info'): # PyPy has different memory management gc.collect() else: # CPython - try to release unused memory ctypes.pythonapi.PyGC_Collect() current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After PyGC_Collect - Memory usage: {current_memory:,.2f} MB") except (AttributeError, TypeError): logger.debug("PyGC_Collect not supported") # 10. Clear Flask-specific caches if applicable if app: try: # Clear Flask caches if they exist for key in list(app.config.get('_cache', {}).keys()): app.config['_cache'].pop(key, None) # Clear Jinja2 template cache if available if hasattr(app, 'jinja_env') and hasattr(app.jinja_env, 'cache'): app.jinja_env.cache.clear() current_memory = process.memory_info().rss / 1024 / 1024 logger.debug(f"After Flask cache clear - Memory usage: {current_memory:,.2f} MB") except (AttributeError, KeyError): logger.debug("No Flask cache to clear") # Final garbage collection pass gc.collect() libc.malloc_trim(0) # Log final memory usage final_memory = process.memory_info().rss / 1024 / 1024 logger.info(f"Memory cleanup completed - Final memory usage: {final_memory:,.2f} MB") return "cleaned"
Perform comprehensive memory cleanup operations and log memory usage at each step with nicely formatted numbers. Args: app: Optional Flask app instance for clearing Flask-specific caches Returns: str: Status message
memory_cleanup
python
dgtlmoon/changedetection.io
changedetectionio/gc_cleanup.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/gc_cleanup.py
Apache-2.0
def element_removal(selectors: List[str], html_content): """Removes elements that match a list of CSS or XPath selectors.""" modified_html = html_content css_selectors = [] xpath_selectors = [] for selector in selectors: if selector.startswith(('xpath:', 'xpath1:', '//')): # Handle XPath selectors separately xpath_selector = selector.removeprefix('xpath:').removeprefix('xpath1:') xpath_selectors.append(xpath_selector) else: # Collect CSS selectors as one "hit", see comment in subtractive_css_selector css_selectors.append(selector.strip().strip(",")) if xpath_selectors: modified_html = subtractive_xpath_selector(xpath_selectors, modified_html) if css_selectors: # Remove duplicates, then combine all CSS selectors into one string, separated by commas # This stops the elements index shifting unique_selectors = list(set(css_selectors)) # Ensure uniqueness combined_css_selector = " , ".join(unique_selectors) modified_html = subtractive_css_selector(combined_css_selector, modified_html) return modified_html
Removes elements that match a list of CSS or XPath selectors.
element_removal
python
dgtlmoon/changedetection.io
changedetectionio/html_tools.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/html_tools.py
Apache-2.0
def workarounds_for_obfuscations(content): """ Some sites are using sneaky tactics to make prices and other information un-renderable by Inscriptis This could go into its own Pip package in the future, for faster updates """ # HomeDepot.com style <span>$<!-- -->90<!-- -->.<!-- -->74</span> # https://github.com/weblyzard/inscriptis/issues/45 if not content: return content content = re.sub('<!--\s+-->', '', content) return content
Some sites are using sneaky tactics to make prices and other information un-renderable by Inscriptis This could go into its own Pip package in the future, for faster updates
workarounds_for_obfuscations
python
dgtlmoon/changedetection.io
changedetectionio/html_tools.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/html_tools.py
Apache-2.0
def queue_notification_for_watch(self, n_object, watch): """ Queue a notification for a watch with full diff rendering and template variables """ from changedetectionio import diff from changedetectionio.notification import default_notification_format_for_watch dates = [] trigger_text = '' now = time.time() if watch: watch_history = watch.history dates = list(watch_history.keys()) trigger_text = watch.get('trigger_text', []) # Add text that was triggered if len(dates): snapshot_contents = watch.get_history_snapshot(dates[-1]) else: snapshot_contents = "No snapshot/history available, the watch should fetch atleast once." # If we ended up here with "System default" if n_object.get('notification_format') == default_notification_format_for_watch: n_object['notification_format'] = self.datastore.data['settings']['application'].get('notification_format') html_colour_enable = False # HTML needs linebreak, but MarkDown and Text can use a linefeed if n_object.get('notification_format') == 'HTML': line_feed_sep = "<br>" # Snapshot will be plaintext on the disk, convert to some kind of HTML snapshot_contents = snapshot_contents.replace('\n', line_feed_sep) elif n_object.get('notification_format') == 'HTML Color': line_feed_sep = "<br>" # Snapshot will be plaintext on the disk, convert to some kind of HTML snapshot_contents = snapshot_contents.replace('\n', line_feed_sep) html_colour_enable = True else: line_feed_sep = "\n" triggered_text = '' if len(trigger_text): from . import html_tools triggered_text = html_tools.get_triggered_text(content=snapshot_contents, trigger_text=trigger_text) if triggered_text: triggered_text = line_feed_sep.join(triggered_text) # Could be called as a 'test notification' with only 1 snapshot available prev_snapshot = "Example text: example test\nExample text: change detection is cool\nExample text: some more examples\n" current_snapshot = "Example text: example test\nExample text: change detection is fantastic\nExample text: even more examples\nExample text: a lot more examples" if len(dates) > 1: prev_snapshot = watch.get_history_snapshot(dates[-2]) current_snapshot = watch.get_history_snapshot(dates[-1]) n_object.update({ 'current_snapshot': snapshot_contents, 'diff': diff.render_diff(prev_snapshot, current_snapshot, line_feed_sep=line_feed_sep, html_colour=html_colour_enable), 'diff_added': diff.render_diff(prev_snapshot, current_snapshot, include_removed=False, line_feed_sep=line_feed_sep), 'diff_full': diff.render_diff(prev_snapshot, current_snapshot, include_equal=True, line_feed_sep=line_feed_sep, html_colour=html_colour_enable), 'diff_patch': diff.render_diff(prev_snapshot, current_snapshot, line_feed_sep=line_feed_sep, patch_format=True), 'diff_removed': diff.render_diff(prev_snapshot, current_snapshot, include_added=False, line_feed_sep=line_feed_sep), 'notification_timestamp': now, 'screenshot': watch.get_screenshot() if watch and watch.get('notification_screenshot') else None, 'triggered_text': triggered_text, 'uuid': watch.get('uuid') if watch else None, 'watch_url': watch.get('url') if watch else None, }) if watch: n_object.update(watch.extra_notification_token_values()) logger.trace(f"Main rendered notification placeholders (diff_added etc) calculated in {time.time()-now:.3f}s") logger.debug("Queued notification for sending") self.notification_q.put(n_object)
Queue a notification for a watch with full diff rendering and template variables
queue_notification_for_watch
python
dgtlmoon/changedetection.io
changedetectionio/notification_service.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/notification_service.py
Apache-2.0
def _check_cascading_vars(self, var_name, watch): """ Check notification variables in cascading priority: Individual watch settings > Tag settings > Global settings """ from changedetectionio.notification import ( default_notification_format_for_watch, default_notification_body, default_notification_title ) # Would be better if this was some kind of Object where Watch can reference the parent datastore etc v = watch.get(var_name) if v and not watch.get('notification_muted'): if var_name == 'notification_format' and v == default_notification_format_for_watch: return self.datastore.data['settings']['application'].get('notification_format') return v tags = self.datastore.get_all_tags_for_watch(uuid=watch.get('uuid')) if tags: for tag_uuid, tag in tags.items(): v = tag.get(var_name) if v and not tag.get('notification_muted'): return v if self.datastore.data['settings']['application'].get(var_name): return self.datastore.data['settings']['application'].get(var_name) # Otherwise could be defaults if var_name == 'notification_format': return default_notification_format_for_watch if var_name == 'notification_body': return default_notification_body if var_name == 'notification_title': return default_notification_title return None
Check notification variables in cascading priority: Individual watch settings > Tag settings > Global settings
_check_cascading_vars
python
dgtlmoon/changedetection.io
changedetectionio/notification_service.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/notification_service.py
Apache-2.0
def send_content_changed_notification(self, watch_uuid): """ Send notification when content changes are detected """ n_object = {} watch = self.datastore.data['watching'].get(watch_uuid) if not watch: return watch_history = watch.history dates = list(watch_history.keys()) # Theoretically it's possible that this could be just 1 long, # - In the case that the timestamp key was not unique if len(dates) == 1: raise ValueError( "History index had 2 or more, but only 1 date loaded, timestamps were not unique? maybe two of the same timestamps got written, needs more delay?" ) # Should be a better parent getter in the model object # Prefer - Individual watch settings > Tag settings > Global settings (in that order) n_object['notification_urls'] = self._check_cascading_vars('notification_urls', watch) n_object['notification_title'] = self._check_cascading_vars('notification_title', watch) n_object['notification_body'] = self._check_cascading_vars('notification_body', watch) n_object['notification_format'] = self._check_cascading_vars('notification_format', watch) # (Individual watch) Only prepare to notify if the rules above matched queued = False if n_object and n_object.get('notification_urls'): queued = True count = watch.get('notification_alert_count', 0) + 1 self.datastore.update_watch(uuid=watch_uuid, update_obj={'notification_alert_count': count}) self.queue_notification_for_watch(n_object=n_object, watch=watch) return queued
Send notification when content changes are detected
send_content_changed_notification
python
dgtlmoon/changedetection.io
changedetectionio/notification_service.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/notification_service.py
Apache-2.0
def send_filter_failure_notification(self, watch_uuid): """ Send notification when CSS/XPath filters fail consecutively """ threshold = self.datastore.data['settings']['application'].get('filter_failure_notification_threshold_attempts') watch = self.datastore.data['watching'].get(watch_uuid) if not watch: return n_object = {'notification_title': 'Changedetection.io - Alert - CSS/xPath filter was not present in the page', 'notification_body': "Your configured CSS/xPath filters of '{}' for {{{{watch_url}}}} did not appear on the page after {} attempts, did the page change layout?\n\nLink: {{{{base_url}}}}/edit/{{{{watch_uuid}}}}\n\nThanks - Your omniscient changedetection.io installation :)\n".format( ", ".join(watch['include_filters']), threshold), 'notification_format': 'text'} if len(watch['notification_urls']): n_object['notification_urls'] = watch['notification_urls'] elif len(self.datastore.data['settings']['application']['notification_urls']): n_object['notification_urls'] = self.datastore.data['settings']['application']['notification_urls'] # Only prepare to notify if the rules above matched if 'notification_urls' in n_object: n_object.update({ 'watch_url': watch['url'], 'uuid': watch_uuid, 'screenshot': None }) self.notification_q.put(n_object) logger.debug(f"Sent filter not found notification for {watch_uuid}") else: logger.debug(f"NOT sending filter not found notification for {watch_uuid} - no notification URLs")
Send notification when CSS/XPath filters fail consecutively
send_filter_failure_notification
python
dgtlmoon/changedetection.io
changedetectionio/notification_service.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/notification_service.py
Apache-2.0
def send_step_failure_notification(self, watch_uuid, step_n): """ Send notification when browser steps fail consecutively """ watch = self.datastore.data['watching'].get(watch_uuid, False) if not watch: return threshold = self.datastore.data['settings']['application'].get('filter_failure_notification_threshold_attempts') n_object = {'notification_title': "Changedetection.io - Alert - Browser step at position {} could not be run".format(step_n+1), 'notification_body': "Your configured browser step at position {} for {{{{watch_url}}}} " "did not appear on the page after {} attempts, did the page change layout? " "Does it need a delay added?\n\nLink: {{{{base_url}}}}/edit/{{{{watch_uuid}}}}\n\n" "Thanks - Your omniscient changedetection.io installation :)\n".format(step_n+1, threshold), 'notification_format': 'text'} if len(watch['notification_urls']): n_object['notification_urls'] = watch['notification_urls'] elif len(self.datastore.data['settings']['application']['notification_urls']): n_object['notification_urls'] = self.datastore.data['settings']['application']['notification_urls'] # Only prepare to notify if the rules above matched if 'notification_urls' in n_object: n_object.update({ 'watch_url': watch['url'], 'uuid': watch_uuid }) self.notification_q.put(n_object) logger.error(f"Sent step not found notification for {watch_uuid}")
Send notification when browser steps fail consecutively
send_step_failure_notification
python
dgtlmoon/changedetection.io
changedetectionio/notification_service.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/notification_service.py
Apache-2.0
def collect_ui_edit_stats_extras(watch): """Collect and combine HTML content from all plugins that implement ui_edit_stats_extras""" extras_content = [] # Get all plugins that implement the ui_edit_stats_extras hook results = plugin_manager.hook.ui_edit_stats_extras(watch=watch) # If we have results, add them to our content if results: for result in results: if result: # Skip empty results extras_content.append(result) return "\n".join(extras_content) if extras_content else ""
Collect and combine HTML content from all plugins that implement ui_edit_stats_extras
collect_ui_edit_stats_extras
python
dgtlmoon/changedetection.io
changedetectionio/pluggy_interface.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/pluggy_interface.py
Apache-2.0
def rehydrate_entity(self, uuid, entity, processor_override=None): """Set the dict back to the dict Watch object""" entity['uuid'] = uuid if processor_override: watch_class = get_custom_watch_obj_for_processor(processor_override) entity['processor']=processor_override else: watch_class = get_custom_watch_obj_for_processor(entity.get('processor')) if entity.get('uuid') != 'text_json_diff': logger.trace(f"Loading Watch object '{watch_class.__module__}.{watch_class.__name__}' for UUID {uuid}") entity = watch_class(datastore_path=self.datastore_path, default=entity) return entity
Set the dict back to the dict Watch object
rehydrate_entity
python
dgtlmoon/changedetection.io
changedetectionio/store.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/store.py
Apache-2.0
def get_preferred_proxy_for_watch(self, uuid): """ Returns the preferred proxy by ID key :param uuid: UUID :return: proxy "key" id """ if self.proxy_list is None: return None # If it's a valid one watch = self.data['watching'].get(uuid) if strtobool(os.getenv('ENABLE_NO_PROXY_OPTION', 'True')) and watch.get('proxy') == "no-proxy": return None if watch.get('proxy') and watch.get('proxy') in list(self.proxy_list.keys()): return watch.get('proxy') # not valid (including None), try the system one else: system_proxy_id = self.data['settings']['requests'].get('proxy') # Is not None and exists if self.proxy_list.get(system_proxy_id): return system_proxy_id # Fallback - Did not resolve anything, or doesnt exist, use the first available if system_proxy_id is None or not self.proxy_list.get(system_proxy_id): first_default = list(self.proxy_list)[0] return first_default return None
Returns the preferred proxy by ID key :param uuid: UUID :return: proxy "key" id
get_preferred_proxy_for_watch
python
dgtlmoon/changedetection.io
changedetectionio/store.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/store.py
Apache-2.0
def get_all_tags_for_watch(self, uuid): """This should be in Watch model but Watch doesn't have access to datastore, not sure how to solve that yet""" watch = self.data['watching'].get(uuid) # Should return a dict of full tag info linked by UUID if watch: return dictfilt(self.__data['settings']['application']['tags'], watch.get('tags', [])) return {}
This should be in Watch model but Watch doesn't have access to datastore, not sure how to solve that yet
get_all_tags_for_watch
python
dgtlmoon/changedetection.io
changedetectionio/store.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/store.py
Apache-2.0
def am_i_inside_time( day_of_week: str, time_str: str, timezone_str: str, duration: int = 15, ) -> bool: """ Determines if the current time falls within a specified time range. Parameters: day_of_week (str): The day of the week (e.g., 'Monday'). time_str (str): The start time in 'HH:MM' format. timezone_str (str): The timezone identifier (e.g., 'Europe/Berlin'). duration (int, optional): The duration of the time range in minutes. Default is 15. Returns: bool: True if the current time is within the time range, False otherwise. """ # Parse the target day of the week try: target_weekday = Weekday[day_of_week.capitalize()] except KeyError: raise ValueError(f"Invalid day_of_week: '{day_of_week}'. Must be a valid weekday name.") # Parse the start time try: target_time = datetime.strptime(time_str, '%H:%M').time() except ValueError: raise ValueError(f"Invalid time_str: '{time_str}'. Must be in 'HH:MM' format.") # Define the timezone try: tz = ZoneInfo(timezone_str) except Exception: raise ValueError(f"Invalid timezone_str: '{timezone_str}'. Must be a valid timezone identifier.") # Get the current time in the specified timezone now_tz = datetime.now(tz) # Check if the current day matches the target day or overlaps due to duration current_weekday = now_tz.weekday() start_datetime_tz = datetime.combine(now_tz.date(), target_time, tzinfo=tz) # Handle previous day's overlap if target_weekday == (current_weekday - 1) % 7: # Calculate start and end times for the overlap from the previous day start_datetime_tz -= timedelta(days=1) end_datetime_tz = start_datetime_tz + timedelta(minutes=duration) if start_datetime_tz <= now_tz < end_datetime_tz: return True # Handle current day's range if target_weekday == current_weekday: end_datetime_tz = start_datetime_tz + timedelta(minutes=duration) if start_datetime_tz <= now_tz < end_datetime_tz: return True # Handle next day's overlap if target_weekday == (current_weekday + 1) % 7: end_datetime_tz = start_datetime_tz + timedelta(minutes=duration) if now_tz < start_datetime_tz and now_tz + timedelta(days=1) < end_datetime_tz: return True return False
Determines if the current time falls within a specified time range. Parameters: day_of_week (str): The day of the week (e.g., 'Monday'). time_str (str): The start time in 'HH:MM' format. timezone_str (str): The timezone identifier (e.g., 'Europe/Berlin'). duration (int, optional): The duration of the time range in minutes. Default is 15. Returns: bool: True if the current time is within the time range, False otherwise.
am_i_inside_time
python
dgtlmoon/changedetection.io
changedetectionio/time_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/time_handler.py
Apache-2.0
def start_async_event_loop(): """Start a dedicated event loop for async workers in a separate thread""" global async_loop logger.info("Starting async event loop for workers") try: # Create a new event loop for this thread async_loop = asyncio.new_event_loop() # Set it as the event loop for this thread asyncio.set_event_loop(async_loop) logger.debug(f"Event loop created and set: {async_loop}") # Run the event loop forever async_loop.run_forever() except Exception as e: logger.error(f"Async event loop error: {e}") finally: # Clean up if async_loop and not async_loop.is_closed(): async_loop.close() async_loop = None logger.info("Async event loop stopped")
Start a dedicated event loop for async workers in a separate thread
start_async_event_loop
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def start_async_workers(n_workers, update_q, notification_q, app, datastore): """Start the async worker management system""" global async_loop_thread, async_loop, running_async_tasks, currently_processing_uuids # Clear any stale UUID tracking state currently_processing_uuids.clear() # Start the event loop in a separate thread async_loop_thread = threading.Thread(target=start_async_event_loop, daemon=True) async_loop_thread.start() # Wait for the loop to be available (with timeout for safety) max_wait_time = 5.0 wait_start = time.time() while async_loop is None and (time.time() - wait_start) < max_wait_time: time.sleep(0.1) if async_loop is None: logger.error("Failed to start async event loop within timeout") return # Additional brief wait to ensure loop is running time.sleep(0.2) # Start async workers logger.info(f"Starting {n_workers} async workers") for i in range(n_workers): try: # Use a factory function to create named worker coroutines def create_named_worker(worker_id): async def named_worker(): task = asyncio.current_task() if task: task.set_name(f"async-worker-{worker_id}") return await start_single_async_worker(worker_id, update_q, notification_q, app, datastore) return named_worker() task_future = asyncio.run_coroutine_threadsafe(create_named_worker(i), async_loop) running_async_tasks.append(task_future) except RuntimeError as e: logger.error(f"Failed to start async worker {i}: {e}") continue
Start the async worker management system
start_async_workers
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
async def start_single_async_worker(worker_id, update_q, notification_q, app, datastore): """Start a single async worker with auto-restart capability""" from changedetectionio.async_update_worker import async_update_worker # Check if we're in pytest environment - if so, be more gentle with logging import os in_pytest = "pytest" in os.sys.modules or "PYTEST_CURRENT_TEST" in os.environ while not app.config.exit.is_set(): try: if not in_pytest: logger.info(f"Starting async worker {worker_id}") await async_update_worker(worker_id, update_q, notification_q, app, datastore) # If we reach here, worker exited cleanly if not in_pytest: logger.info(f"Async worker {worker_id} exited cleanly") break except asyncio.CancelledError: # Task was cancelled (normal shutdown) if not in_pytest: logger.info(f"Async worker {worker_id} cancelled") break except Exception as e: logger.error(f"Async worker {worker_id} crashed: {e}") if not in_pytest: logger.info(f"Restarting async worker {worker_id} in 5 seconds...") await asyncio.sleep(5) if not in_pytest: logger.info(f"Async worker {worker_id} shutdown complete")
Start a single async worker with auto-restart capability
start_single_async_worker
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def add_worker(update_q, notification_q, app, datastore): """Add a new async worker (for dynamic scaling)""" global running_async_tasks if not async_loop: logger.error("Async loop not running, cannot add worker") return False worker_id = len(running_async_tasks) logger.info(f"Adding async worker {worker_id}") task_future = asyncio.run_coroutine_threadsafe( start_single_async_worker(worker_id, update_q, notification_q, app, datastore), async_loop ) running_async_tasks.append(task_future) return True
Add a new async worker (for dynamic scaling)
add_worker
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def remove_worker(): """Remove an async worker (for dynamic scaling)""" global running_async_tasks if not running_async_tasks: return False # Cancel the last worker task_future = running_async_tasks.pop() task_future.cancel() logger.info(f"Removed async worker, {len(running_async_tasks)} workers remaining") return True
Remove an async worker (for dynamic scaling)
remove_worker
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def set_uuid_processing(uuid, processing=True): """Mark a UUID as being processed or completed""" global currently_processing_uuids if processing: currently_processing_uuids.add(uuid) logger.debug(f"Started processing UUID: {uuid}") else: currently_processing_uuids.discard(uuid) logger.debug(f"Finished processing UUID: {uuid}")
Mark a UUID as being processed or completed
set_uuid_processing
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def queue_item_async_safe(update_q, item): """Queue an item for async queue processing""" if async_loop and not async_loop.is_closed(): try: # For async queue, schedule the put operation asyncio.run_coroutine_threadsafe(update_q.put(item), async_loop) except RuntimeError as e: logger.error(f"Failed to queue item: {e}") else: logger.error("Async loop not available or closed for queueing item")
Queue an item for async queue processing
queue_item_async_safe
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def shutdown_workers(): """Shutdown all async workers fast and aggressively""" global async_loop, async_loop_thread, running_async_tasks # Check if we're in pytest environment - if so, be more gentle with logging import os in_pytest = "pytest" in os.sys.modules or "PYTEST_CURRENT_TEST" in os.environ if not in_pytest: logger.info("Fast shutdown of async workers initiated...") # Cancel all async tasks immediately for task_future in running_async_tasks: if not task_future.done(): task_future.cancel() # Stop the async event loop immediately if async_loop and not async_loop.is_closed(): try: async_loop.call_soon_threadsafe(async_loop.stop) except RuntimeError: # Loop might already be stopped pass running_async_tasks.clear() async_loop = None # Give async thread minimal time to finish, then continue if async_loop_thread and async_loop_thread.is_alive(): async_loop_thread.join(timeout=1.0) # Only 1 second timeout if async_loop_thread.is_alive() and not in_pytest: logger.info("Async thread still running after timeout - continuing with shutdown") async_loop_thread = None if not in_pytest: logger.info("Async workers fast shutdown complete")
Shutdown all async workers fast and aggressively
shutdown_workers
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def adjust_async_worker_count(new_count, update_q=None, notification_q=None, app=None, datastore=None): """ Dynamically adjust the number of async workers. Args: new_count: Target number of workers update_q, notification_q, app, datastore: Required for adding new workers Returns: dict: Status of the adjustment operation """ global running_async_tasks current_count = get_worker_count() if new_count == current_count: return { 'status': 'no_change', 'message': f'Worker count already at {current_count}', 'current_count': current_count } if new_count > current_count: # Add workers workers_to_add = new_count - current_count logger.info(f"Adding {workers_to_add} async workers (from {current_count} to {new_count})") if not all([update_q, notification_q, app, datastore]): return { 'status': 'error', 'message': 'Missing required parameters to add workers', 'current_count': current_count } for i in range(workers_to_add): worker_id = len(running_async_tasks) task_future = asyncio.run_coroutine_threadsafe( start_single_async_worker(worker_id, update_q, notification_q, app, datastore), async_loop ) running_async_tasks.append(task_future) return { 'status': 'success', 'message': f'Added {workers_to_add} workers', 'previous_count': current_count, 'current_count': new_count } else: # Remove workers workers_to_remove = current_count - new_count logger.info(f"Removing {workers_to_remove} async workers (from {current_count} to {new_count})") removed_count = 0 for _ in range(workers_to_remove): if running_async_tasks: task_future = running_async_tasks.pop() task_future.cancel() # Wait for the task to actually stop try: task_future.result(timeout=5) # 5 second timeout except Exception: pass # Task was cancelled, which is expected removed_count += 1 return { 'status': 'success', 'message': f'Removed {removed_count} workers', 'previous_count': current_count, 'current_count': current_count - removed_count }
Dynamically adjust the number of async workers. Args: new_count: Target number of workers update_q, notification_q, app, datastore: Required for adding new workers Returns: dict: Status of the adjustment operation
adjust_async_worker_count
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def get_worker_status(): """Get status information about async workers""" return { 'worker_type': 'async', 'worker_count': get_worker_count(), 'running_uuids': get_running_uuids(), 'async_loop_running': async_loop is not None, }
Get status information about async workers
get_worker_status
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def check_worker_health(expected_count, update_q=None, notification_q=None, app=None, datastore=None): """ Check if the expected number of async workers are running and restart any missing ones. Args: expected_count: Expected number of workers update_q, notification_q, app, datastore: Required for restarting workers Returns: dict: Health check results """ global running_async_tasks current_count = get_worker_count() if current_count == expected_count: return { 'status': 'healthy', 'expected_count': expected_count, 'actual_count': current_count, 'message': f'All {expected_count} async workers running' } # Check for crashed async workers dead_workers = [] alive_count = 0 for i, task_future in enumerate(running_async_tasks[:]): if task_future.done(): try: result = task_future.result() dead_workers.append(i) logger.warning(f"Async worker {i} completed unexpectedly") except Exception as e: dead_workers.append(i) logger.error(f"Async worker {i} crashed: {e}") else: alive_count += 1 # Remove dead workers from tracking for i in reversed(dead_workers): if i < len(running_async_tasks): running_async_tasks.pop(i) missing_workers = expected_count - alive_count restarted_count = 0 if missing_workers > 0 and all([update_q, notification_q, app, datastore]): logger.info(f"Restarting {missing_workers} crashed async workers") for i in range(missing_workers): worker_id = alive_count + i try: task_future = asyncio.run_coroutine_threadsafe( start_single_async_worker(worker_id, update_q, notification_q, app, datastore), async_loop ) running_async_tasks.append(task_future) restarted_count += 1 except Exception as e: logger.error(f"Failed to restart worker {worker_id}: {e}") return { 'status': 'repaired' if restarted_count > 0 else 'degraded', 'expected_count': expected_count, 'actual_count': alive_count, 'dead_workers': len(dead_workers), 'restarted_workers': restarted_count, 'message': f'Found {len(dead_workers)} dead workers, restarted {restarted_count}' }
Check if the expected number of async workers are running and restart any missing ones. Args: expected_count: Expected number of workers update_q, notification_q, app, datastore: Required for restarting workers Returns: dict: Health check results
check_worker_health
python
dgtlmoon/changedetection.io
changedetectionio/worker_handler.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/worker_handler.py
Apache-2.0
def post(self): """ @api {post} /api/v1/import Import a list of watched URLs @apiDescription Accepts a line-feed separated list of URLs to import, additionally with ?tag_uuids=(tag id), ?tag=(name), ?proxy={key}, ?dedupe=true (default true) one URL per line. @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/import --data-binary @list-of-sites.txt -H"x-api-key:8a111a21bc2f8f1dd9b9353bbd46049a" @apiName Import @apiGroup Watch @apiSuccess (200) {List} OK List of watch UUIDs added @apiSuccess (500) {String} ERR Some other error """ extras = {} if request.args.get('proxy'): plist = self.datastore.proxy_list if not request.args.get('proxy') in plist: return "Invalid proxy choice, currently supported proxies are '{}'".format(', '.join(plist)), 400 else: extras['proxy'] = request.args.get('proxy') dedupe = strtobool(request.args.get('dedupe', 'true')) tags = request.args.get('tag') tag_uuids = request.args.get('tag_uuids') if tag_uuids: tag_uuids = tag_uuids.split(',') urls = request.get_data().decode('utf8').splitlines() added = [] allow_simplehost = not strtobool(os.getenv('BLOCK_SIMPLEHOSTS', 'False')) for url in urls: url = url.strip() if not len(url): continue # If hosts that only contain alphanumerics are allowed ("localhost" for example) if not validators.url(url, simple_host=allow_simplehost): return f"Invalid or unsupported URL - {url}", 400 if dedupe and self.datastore.url_exists(url): continue new_uuid = self.datastore.add_watch(url=url, extras=extras, tag=tags, tag_uuids=tag_uuids) added.append(new_uuid) return added
@api {post} /api/v1/import Import a list of watched URLs @apiDescription Accepts a line-feed separated list of URLs to import, additionally with ?tag_uuids=(tag id), ?tag=(name), ?proxy={key}, ?dedupe=true (default true) one URL per line. @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/import --data-binary @list-of-sites.txt -H"x-api-key:8a111a21bc2f8f1dd9b9353bbd46049a" @apiName Import @apiGroup Watch @apiSuccess (200) {List} OK List of watch UUIDs added @apiSuccess (500) {String} ERR Some other error
post
python
dgtlmoon/changedetection.io
changedetectionio/api/Import.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Import.py
Apache-2.0
def get(self): """ @api {get} /api/v1/notifications Return Notification URL List @apiDescription Return the Notification URL List from the configuration @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/notifications -H"x-api-key:813031b16330fe25e3780cf0325daa45" HTTP/1.0 200 { 'notification_urls': ["notification-urls-list"] } @apiName Get @apiGroup Notifications """ notification_urls = self.datastore.data.get('settings', {}).get('application', {}).get('notification_urls', []) return { 'notification_urls': notification_urls, }, 200
@api {get} /api/v1/notifications Return Notification URL List @apiDescription Return the Notification URL List from the configuration @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/notifications -H"x-api-key:813031b16330fe25e3780cf0325daa45" HTTP/1.0 200 { 'notification_urls': ["notification-urls-list"] } @apiName Get @apiGroup Notifications
get
python
dgtlmoon/changedetection.io
changedetectionio/api/Notifications.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Notifications.py
Apache-2.0
def post(self): """ @api {post} /api/v1/notifications Create Notification URLs @apiDescription Add one or more notification URLs from the configuration @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/notifications/batch -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"notification_urls": ["url1", "url2"]}' @apiName CreateBatch @apiGroup Notifications @apiSuccess (201) {Object[]} notification_urls List of added notification URLs @apiError (400) {String} Invalid input """ json_data = request.get_json() notification_urls = json_data.get("notification_urls", []) from wtforms import ValidationError try: validate_notification_urls(notification_urls) except ValidationError as e: return str(e), 400 added_urls = [] for url in notification_urls: clean_url = url.strip() added_url = self.datastore.add_notification_url(clean_url) if added_url: added_urls.append(added_url) if not added_urls: return "No valid notification URLs were added", 400 return {'notification_urls': added_urls}, 201
@api {post} /api/v1/notifications Create Notification URLs @apiDescription Add one or more notification URLs from the configuration @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/notifications/batch -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"notification_urls": ["url1", "url2"]}' @apiName CreateBatch @apiGroup Notifications @apiSuccess (201) {Object[]} notification_urls List of added notification URLs @apiError (400) {String} Invalid input
post
python
dgtlmoon/changedetection.io
changedetectionio/api/Notifications.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Notifications.py
Apache-2.0
def put(self): """ @api {put} /api/v1/notifications Replace Notification URLs @apiDescription Replace all notification URLs with the provided list (can be empty) @apiExample {curl} Example usage: curl -X PUT http://localhost:5000/api/v1/notifications -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"notification_urls": ["url1", "url2"]}' @apiName Replace @apiGroup Notifications @apiSuccess (200) {Object[]} notification_urls List of current notification URLs @apiError (400) {String} Invalid input """ json_data = request.get_json() notification_urls = json_data.get("notification_urls", []) from wtforms import ValidationError try: validate_notification_urls(notification_urls) except ValidationError as e: return str(e), 400 if not isinstance(notification_urls, list): return "Invalid input format", 400 clean_urls = [url.strip() for url in notification_urls if isinstance(url, str)] self.datastore.data['settings']['application']['notification_urls'] = clean_urls self.datastore.needs_write = True return {'notification_urls': clean_urls}, 200
@api {put} /api/v1/notifications Replace Notification URLs @apiDescription Replace all notification URLs with the provided list (can be empty) @apiExample {curl} Example usage: curl -X PUT http://localhost:5000/api/v1/notifications -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"notification_urls": ["url1", "url2"]}' @apiName Replace @apiGroup Notifications @apiSuccess (200) {Object[]} notification_urls List of current notification URLs @apiError (400) {String} Invalid input
put
python
dgtlmoon/changedetection.io
changedetectionio/api/Notifications.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Notifications.py
Apache-2.0
def delete(self): """ @api {delete} /api/v1/notifications Delete Notification URLs @apiDescription Deletes one or more notification URLs from the configuration @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/notifications -X DELETE -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"notification_urls": ["url1", "url2"]}' @apiParam {String[]} notification_urls The notification URLs to delete. @apiName Delete @apiGroup Notifications @apiSuccess (204) {String} OK Deleted @apiError (400) {String} No matching notification URLs found. """ json_data = request.get_json() urls_to_delete = json_data.get("notification_urls", []) if not isinstance(urls_to_delete, list): abort(400, message="Expected a list of notification URLs.") notification_urls = self.datastore.data['settings']['application'].get('notification_urls', []) deleted = [] for url in urls_to_delete: clean_url = url.strip() if clean_url in notification_urls: notification_urls.remove(clean_url) deleted.append(clean_url) if not deleted: abort(400, message="No matching notification URLs found.") self.datastore.data['settings']['application']['notification_urls'] = notification_urls self.datastore.needs_write = True return 'OK', 204
@api {delete} /api/v1/notifications Delete Notification URLs @apiDescription Deletes one or more notification URLs from the configuration @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/notifications -X DELETE -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"notification_urls": ["url1", "url2"]}' @apiParam {String[]} notification_urls The notification URLs to delete. @apiName Delete @apiGroup Notifications @apiSuccess (204) {String} OK Deleted @apiError (400) {String} No matching notification URLs found.
delete
python
dgtlmoon/changedetection.io
changedetectionio/api/Notifications.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Notifications.py
Apache-2.0
def get(self): """ @api {get} /api/v1/systeminfo Return system info @apiDescription Return some info about the current system state @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/systeminfo -H"x-api-key:813031b16330fe25e3780cf0325daa45" HTTP/1.0 200 { 'queue_size': 10 , 'overdue_watches': ["watch-uuid-list"], 'uptime': 38344.55, 'watch_count': 800, 'version': "0.40.1" } @apiName Get Info @apiGroup System Information """ import time overdue_watches = [] # Check all watches and report which have not been checked but should have been for uuid, watch in self.datastore.data.get('watching', {}).items(): # see if now - last_checked is greater than the time that should have been # this is not super accurate (maybe they just edited it) but better than nothing t = watch.threshold_seconds() if not t: # Use the system wide default t = self.datastore.threshold_seconds time_since_check = time.time() - watch.get('last_checked') # Allow 5 minutes of grace time before we decide it's overdue if time_since_check - (5 * 60) > t: overdue_watches.append(uuid) from changedetectionio import __version__ as main_version return { 'queue_size': self.update_q.qsize(), 'overdue_watches': overdue_watches, 'uptime': round(time.time() - self.datastore.start_time, 2), 'watch_count': len(self.datastore.data.get('watching', {})), 'version': main_version }, 200
@api {get} /api/v1/systeminfo Return system info @apiDescription Return some info about the current system state @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/systeminfo -H"x-api-key:813031b16330fe25e3780cf0325daa45" HTTP/1.0 200 { 'queue_size': 10 , 'overdue_watches': ["watch-uuid-list"], 'uptime': 38344.55, 'watch_count': 800, 'version': "0.40.1" } @apiName Get Info @apiGroup System Information
get
python
dgtlmoon/changedetection.io
changedetectionio/api/SystemInfo.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/SystemInfo.py
Apache-2.0
def get(self, uuid): """ @api {get} /api/v1/tag/:uuid Single tag - get data or toggle notification muting. @apiDescription Retrieve tag information and set notification_muted status @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/tag/cc0cfffa-f449-477b-83ea-0caafd1dc091 -H"x-api-key:813031b16330fe25e3780cf0325daa45" curl "http://localhost:5000/api/v1/tag/cc0cfffa-f449-477b-83ea-0caafd1dc091?muted=muted" -H"x-api-key:813031b16330fe25e3780cf0325daa45" @apiName Tag @apiGroup Tag @apiParam {uuid} uuid Tag unique ID. @apiQuery {String} [muted] =`muted` or =`unmuted` , Sets the MUTE NOTIFICATIONS state @apiSuccess (200) {String} OK When muted operation OR full JSON object of the tag @apiSuccess (200) {JSON} TagJSON JSON Full JSON object of the tag """ from copy import deepcopy tag = deepcopy(self.datastore.data['settings']['application']['tags'].get(uuid)) if not tag: abort(404, message=f'No tag exists with the UUID of {uuid}') if request.args.get('muted', '') == 'muted': self.datastore.data['settings']['application']['tags'][uuid]['notification_muted'] = True return "OK", 200 elif request.args.get('muted', '') == 'unmuted': self.datastore.data['settings']['application']['tags'][uuid]['notification_muted'] = False return "OK", 200 return tag
@api {get} /api/v1/tag/:uuid Single tag - get data or toggle notification muting. @apiDescription Retrieve tag information and set notification_muted status @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/tag/cc0cfffa-f449-477b-83ea-0caafd1dc091 -H"x-api-key:813031b16330fe25e3780cf0325daa45" curl "http://localhost:5000/api/v1/tag/cc0cfffa-f449-477b-83ea-0caafd1dc091?muted=muted" -H"x-api-key:813031b16330fe25e3780cf0325daa45" @apiName Tag @apiGroup Tag @apiParam {uuid} uuid Tag unique ID. @apiQuery {String} [muted] =`muted` or =`unmuted` , Sets the MUTE NOTIFICATIONS state @apiSuccess (200) {String} OK When muted operation OR full JSON object of the tag @apiSuccess (200) {JSON} TagJSON JSON Full JSON object of the tag
get
python
dgtlmoon/changedetection.io
changedetectionio/api/Tags.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Tags.py
Apache-2.0
def delete(self, uuid): """ @api {delete} /api/v1/tag/:uuid Delete a tag and remove it from all watches @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/tag/cc0cfffa-f449-477b-83ea-0caafd1dc091 -X DELETE -H"x-api-key:813031b16330fe25e3780cf0325daa45" @apiParam {uuid} uuid Tag unique ID. @apiName DeleteTag @apiGroup Tag @apiSuccess (200) {String} OK Was deleted """ if not self.datastore.data['settings']['application']['tags'].get(uuid): abort(400, message='No tag exists with the UUID of {}'.format(uuid)) # Delete the tag, and any tag reference del self.datastore.data['settings']['application']['tags'][uuid] # Remove tag from all watches for watch_uuid, watch in self.datastore.data['watching'].items(): if watch.get('tags') and uuid in watch['tags']: watch['tags'].remove(uuid) return 'OK', 204
@api {delete} /api/v1/tag/:uuid Delete a tag and remove it from all watches @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/tag/cc0cfffa-f449-477b-83ea-0caafd1dc091 -X DELETE -H"x-api-key:813031b16330fe25e3780cf0325daa45" @apiParam {uuid} uuid Tag unique ID. @apiName DeleteTag @apiGroup Tag @apiSuccess (200) {String} OK Was deleted
delete
python
dgtlmoon/changedetection.io
changedetectionio/api/Tags.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Tags.py
Apache-2.0
def put(self, uuid): """ @api {put} /api/v1/tag/:uuid Update tag information @apiExample {curl} Example usage: Update (PUT) curl http://localhost:5000/api/v1/tag/cc0cfffa-f449-477b-83ea-0caafd1dc091 -X PUT -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"title": "New Tag Title"}' @apiDescription Updates an existing tag using JSON @apiParam {uuid} uuid Tag unique ID. @apiName UpdateTag @apiGroup Tag @apiSuccess (200) {String} OK Was updated @apiSuccess (500) {String} ERR Some other error """ tag = self.datastore.data['settings']['application']['tags'].get(uuid) if not tag: abort(404, message='No tag exists with the UUID of {}'.format(uuid)) tag.update(request.json) self.datastore.needs_write_urgent = True return "OK", 200
@api {put} /api/v1/tag/:uuid Update tag information @apiExample {curl} Example usage: Update (PUT) curl http://localhost:5000/api/v1/tag/cc0cfffa-f449-477b-83ea-0caafd1dc091 -X PUT -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"title": "New Tag Title"}' @apiDescription Updates an existing tag using JSON @apiParam {uuid} uuid Tag unique ID. @apiName UpdateTag @apiGroup Tag @apiSuccess (200) {String} OK Was updated @apiSuccess (500) {String} ERR Some other error
put
python
dgtlmoon/changedetection.io
changedetectionio/api/Tags.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Tags.py
Apache-2.0
def post(self): """ @api {post} /api/v1/watch Create a single tag @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"name": "Work related"}' @apiName Create @apiGroup Tag @apiSuccess (200) {String} OK Was created @apiSuccess (500) {String} ERR Some other error """ json_data = request.get_json() title = json_data.get("title",'').strip() new_uuid = self.datastore.add_tag(title=title) if new_uuid: return {'uuid': new_uuid}, 201 else: return "Invalid or unsupported tag", 400
@api {post} /api/v1/watch Create a single tag @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"name": "Work related"}' @apiName Create @apiGroup Tag @apiSuccess (200) {String} OK Was created @apiSuccess (500) {String} ERR Some other error
post
python
dgtlmoon/changedetection.io
changedetectionio/api/Tags.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Tags.py
Apache-2.0
def get(self): """ @api {get} /api/v1/tags List tags @apiDescription Return list of available tags @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/tags -H"x-api-key:813031b16330fe25e3780cf0325daa45" { "cc0cfffa-f449-477b-83ea-0caafd1dc091": { "title": "Tech News", "notification_muted": false, "date_created": 1677103794 }, "e6f5fd5c-dbfe-468b-b8f3-f9d6ff5ad69b": { "title": "Shopping", "notification_muted": true, "date_created": 1676662819 } } @apiName ListTags @apiGroup Tag Management @apiSuccess (200) {String} OK JSON dict """ result = {} for uuid, tag in self.datastore.data['settings']['application']['tags'].items(): result[uuid] = { 'date_created': tag.get('date_created', 0), 'notification_muted': tag.get('notification_muted', False), 'title': tag.get('title', ''), 'uuid': tag.get('uuid') } return result, 200
@api {get} /api/v1/tags List tags @apiDescription Return list of available tags @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/tags -H"x-api-key:813031b16330fe25e3780cf0325daa45" { "cc0cfffa-f449-477b-83ea-0caafd1dc091": { "title": "Tech News", "notification_muted": false, "date_created": 1677103794 }, "e6f5fd5c-dbfe-468b-b8f3-f9d6ff5ad69b": { "title": "Shopping", "notification_muted": true, "date_created": 1676662819 } } @apiName ListTags @apiGroup Tag Management @apiSuccess (200) {String} OK JSON dict
get
python
dgtlmoon/changedetection.io
changedetectionio/api/Tags.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Tags.py
Apache-2.0
def get(self, uuid): """ @api {get} /api/v1/watch/:uuid Single watch - get data, recheck, pause, mute. @apiDescription Retrieve watch information and set muted/paused status @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091 -H"x-api-key:813031b16330fe25e3780cf0325daa45" curl "http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091?muted=unmuted" -H"x-api-key:813031b16330fe25e3780cf0325daa45" curl "http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091?paused=unpaused" -H"x-api-key:813031b16330fe25e3780cf0325daa45" @apiName Watch @apiGroup Watch @apiParam {uuid} uuid Watch unique ID. @apiQuery {Boolean} [recheck] Recheck this watch `recheck=1` @apiQuery {String} [paused] =`paused` or =`unpaused` , Sets the PAUSED state @apiQuery {String} [muted] =`muted` or =`unmuted` , Sets the MUTE NOTIFICATIONS state @apiSuccess (200) {String} OK When paused/muted/recheck operation OR full JSON object of the watch @apiSuccess (200) {JSON} WatchJSON JSON Full JSON object of the watch """ from copy import deepcopy watch = deepcopy(self.datastore.data['watching'].get(uuid)) if not watch: abort(404, message='No watch exists with the UUID of {}'.format(uuid)) if request.args.get('recheck'): worker_handler.queue_item_async_safe(self.update_q, queuedWatchMetaData.PrioritizedItem(priority=1, item={'uuid': uuid})) return "OK", 200 if request.args.get('paused', '') == 'paused': self.datastore.data['watching'].get(uuid).pause() return "OK", 200 elif request.args.get('paused', '') == 'unpaused': self.datastore.data['watching'].get(uuid).unpause() return "OK", 200 if request.args.get('muted', '') == 'muted': self.datastore.data['watching'].get(uuid).mute() return "OK", 200 elif request.args.get('muted', '') == 'unmuted': self.datastore.data['watching'].get(uuid).unmute() return "OK", 200 # Return without history, get that via another API call # Properties are not returned as a JSON, so add the required props manually watch['history_n'] = watch.history_n # attr .last_changed will check for the last written text snapshot on change watch['last_changed'] = watch.last_changed watch['viewed'] = watch.viewed return watch
@api {get} /api/v1/watch/:uuid Single watch - get data, recheck, pause, mute. @apiDescription Retrieve watch information and set muted/paused status @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091 -H"x-api-key:813031b16330fe25e3780cf0325daa45" curl "http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091?muted=unmuted" -H"x-api-key:813031b16330fe25e3780cf0325daa45" curl "http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091?paused=unpaused" -H"x-api-key:813031b16330fe25e3780cf0325daa45" @apiName Watch @apiGroup Watch @apiParam {uuid} uuid Watch unique ID. @apiQuery {Boolean} [recheck] Recheck this watch `recheck=1` @apiQuery {String} [paused] =`paused` or =`unpaused` , Sets the PAUSED state @apiQuery {String} [muted] =`muted` or =`unmuted` , Sets the MUTE NOTIFICATIONS state @apiSuccess (200) {String} OK When paused/muted/recheck operation OR full JSON object of the watch @apiSuccess (200) {JSON} WatchJSON JSON Full JSON object of the watch
get
python
dgtlmoon/changedetection.io
changedetectionio/api/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Watch.py
Apache-2.0
def delete(self, uuid): """ @api {delete} /api/v1/watch/:uuid Delete a watch and related history @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091 -X DELETE -H"x-api-key:813031b16330fe25e3780cf0325daa45" @apiParam {uuid} uuid Watch unique ID. @apiName Delete @apiGroup Watch @apiSuccess (200) {String} OK Was deleted """ if not self.datastore.data['watching'].get(uuid): abort(400, message='No watch exists with the UUID of {}'.format(uuid)) self.datastore.delete(uuid) return 'OK', 204
@api {delete} /api/v1/watch/:uuid Delete a watch and related history @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091 -X DELETE -H"x-api-key:813031b16330fe25e3780cf0325daa45" @apiParam {uuid} uuid Watch unique ID. @apiName Delete @apiGroup Watch @apiSuccess (200) {String} OK Was deleted
delete
python
dgtlmoon/changedetection.io
changedetectionio/api/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Watch.py
Apache-2.0
def put(self, uuid): """ @api {put} /api/v1/watch/:uuid Update watch information @apiExample {curl} Example usage: Update (PUT) curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091 -X PUT -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"url": "https://my-nice.com" , "tag": "new list"}' @apiDescription Updates an existing watch using JSON, accepts the same structure as returned in <a href="#api-Watch-Watch">get single watch information</a> @apiParam {uuid} uuid Watch unique ID. @apiName Update a watch @apiGroup Watch @apiSuccess (200) {String} OK Was updated @apiSuccess (500) {String} ERR Some other error """ watch = self.datastore.data['watching'].get(uuid) if not watch: abort(404, message='No watch exists with the UUID of {}'.format(uuid)) if request.json.get('proxy'): plist = self.datastore.proxy_list if not request.json.get('proxy') in plist: return "Invalid proxy choice, currently supported proxies are '{}'".format(', '.join(plist)), 400 watch.update(request.json) return "OK", 200
@api {put} /api/v1/watch/:uuid Update watch information @apiExample {curl} Example usage: Update (PUT) curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091 -X PUT -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"url": "https://my-nice.com" , "tag": "new list"}' @apiDescription Updates an existing watch using JSON, accepts the same structure as returned in <a href="#api-Watch-Watch">get single watch information</a> @apiParam {uuid} uuid Watch unique ID. @apiName Update a watch @apiGroup Watch @apiSuccess (200) {String} OK Was updated @apiSuccess (500) {String} ERR Some other error
put
python
dgtlmoon/changedetection.io
changedetectionio/api/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Watch.py
Apache-2.0
def get(self, uuid): """ @api {get} /api/v1/watch/<string:uuid>/history Get a list of all historical snapshots available for a watch @apiDescription Requires `uuid`, returns list @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091/history -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" { "1676649279": "/tmp/data/6a4b7d5c-fee4-4616-9f43-4ac97046b595/cb7e9be8258368262246910e6a2a4c30.txt", "1677092785": "/tmp/data/6a4b7d5c-fee4-4616-9f43-4ac97046b595/e20db368d6fc633e34f559ff67bb4044.txt", "1677103794": "/tmp/data/6a4b7d5c-fee4-4616-9f43-4ac97046b595/02efdd37dacdae96554a8cc85dc9c945.txt" } @apiName Get list of available stored snapshots for watch @apiGroup Watch History @apiSuccess (200) {String} OK @apiSuccess (404) {String} ERR Not found """ watch = self.datastore.data['watching'].get(uuid) if not watch: abort(404, message='No watch exists with the UUID of {}'.format(uuid)) return watch.history, 200
@api {get} /api/v1/watch/<string:uuid>/history Get a list of all historical snapshots available for a watch @apiDescription Requires `uuid`, returns list @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091/history -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" { "1676649279": "/tmp/data/6a4b7d5c-fee4-4616-9f43-4ac97046b595/cb7e9be8258368262246910e6a2a4c30.txt", "1677092785": "/tmp/data/6a4b7d5c-fee4-4616-9f43-4ac97046b595/e20db368d6fc633e34f559ff67bb4044.txt", "1677103794": "/tmp/data/6a4b7d5c-fee4-4616-9f43-4ac97046b595/02efdd37dacdae96554a8cc85dc9c945.txt" } @apiName Get list of available stored snapshots for watch @apiGroup Watch History @apiSuccess (200) {String} OK @apiSuccess (404) {String} ERR Not found
get
python
dgtlmoon/changedetection.io
changedetectionio/api/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Watch.py
Apache-2.0
def get(self, uuid, timestamp): """ @api {get} /api/v1/watch/<string:uuid>/history/<int:timestamp> Get single snapshot from watch @apiDescription Requires watch `uuid` and `timestamp`. `timestamp` of "`latest`" for latest available snapshot, or <a href="#api-Watch_History-Get_list_of_available_stored_snapshots_for_watch">use the list returned here</a> @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091/history/1677092977 -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" @apiName Get single snapshot content @apiGroup Watch History @apiParam {String} [html] Optional Set to =1 to return the last HTML (only stores last 2 snapshots, use `latest` as timestamp) @apiSuccess (200) {String} OK @apiSuccess (404) {String} ERR Not found """ watch = self.datastore.data['watching'].get(uuid) if not watch: abort(404, message=f"No watch exists with the UUID of {uuid}") if not len(watch.history): abort(404, message=f"Watch found but no history exists for the UUID {uuid}") if timestamp == 'latest': timestamp = list(watch.history.keys())[-1] if request.args.get('html'): content = watch.get_fetched_html(timestamp) if content: response = make_response(content, 200) response.mimetype = "text/html" else: response = make_response("No content found", 404) response.mimetype = "text/plain" else: content = watch.get_history_snapshot(timestamp) response = make_response(content, 200) response.mimetype = "text/plain" return response
@api {get} /api/v1/watch/<string:uuid>/history/<int:timestamp> Get single snapshot from watch @apiDescription Requires watch `uuid` and `timestamp`. `timestamp` of "`latest`" for latest available snapshot, or <a href="#api-Watch_History-Get_list_of_available_stored_snapshots_for_watch">use the list returned here</a> @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch/cc0cfffa-f449-477b-83ea-0caafd1dc091/history/1677092977 -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" @apiName Get single snapshot content @apiGroup Watch History @apiParam {String} [html] Optional Set to =1 to return the last HTML (only stores last 2 snapshots, use `latest` as timestamp) @apiSuccess (200) {String} OK @apiSuccess (404) {String} ERR Not found
get
python
dgtlmoon/changedetection.io
changedetectionio/api/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Watch.py
Apache-2.0
def post(self): """ @api {post} /api/v1/watch Create a single watch @apiDescription Requires atleast `url` set, can accept the same structure as <a href="#api-Watch-Watch">get single watch information</a> to create. @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"url": "https://my-nice.com" , "tag": "nice list"}' @apiName Create @apiGroup Watch @apiSuccess (200) {String} OK Was created @apiSuccess (500) {String} ERR Some other error """ json_data = request.get_json() url = json_data['url'].strip() # If hosts that only contain alphanumerics are allowed ("localhost" for example) allow_simplehost = not strtobool(os.getenv('BLOCK_SIMPLEHOSTS', 'False')) if not validators.url(url, simple_host=allow_simplehost): return "Invalid or unsupported URL", 400 if json_data.get('proxy'): plist = self.datastore.proxy_list if not json_data.get('proxy') in plist: return "Invalid proxy choice, currently supported proxies are '{}'".format(', '.join(plist)), 400 extras = copy.deepcopy(json_data) # Because we renamed 'tag' to 'tags' but don't want to change the API (can do this in v2 of the API) tags = None if extras.get('tag'): tags = extras.get('tag') del extras['tag'] del extras['url'] new_uuid = self.datastore.add_watch(url=url, extras=extras, tag=tags) if new_uuid: worker_handler.queue_item_async_safe(self.update_q, queuedWatchMetaData.PrioritizedItem(priority=1, item={'uuid': new_uuid})) return {'uuid': new_uuid}, 201 else: return "Invalid or unsupported URL", 400
@api {post} /api/v1/watch Create a single watch @apiDescription Requires atleast `url` set, can accept the same structure as <a href="#api-Watch-Watch">get single watch information</a> to create. @apiExample {curl} Example usage: curl http://localhost:5000/api/v1/watch -H"x-api-key:813031b16330fe25e3780cf0325daa45" -H "Content-Type: application/json" -d '{"url": "https://my-nice.com" , "tag": "nice list"}' @apiName Create @apiGroup Watch @apiSuccess (200) {String} OK Was created @apiSuccess (500) {String} ERR Some other error
post
python
dgtlmoon/changedetection.io
changedetectionio/api/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/api/Watch.py
Apache-2.0
async def action_remove_elements(self, selector, value): """Removes all elements matching the given selector from the DOM.""" if not selector: return await self.page.locator(selector).evaluate_all("els => els.forEach(el => el.remove())")
Removes all elements matching the given selector from the DOM.
action_remove_elements
python
dgtlmoon/changedetection.io
changedetectionio/blueprint/browser_steps/browser_steps.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/blueprint/browser_steps/browser_steps.py
Apache-2.0
async def action_make_all_child_elements_visible(self, selector, value): """Recursively makes all child elements inside the given selector fully visible.""" if not selector: return await self.page.locator(selector).locator("*").evaluate_all(""" els => els.forEach(el => { el.style.display = 'block'; // Forces it to be displayed el.style.visibility = 'visible'; // Ensures it's not hidden el.style.opacity = '1'; // Fully opaque el.style.position = 'relative'; // Avoids 'absolute' hiding el.style.height = 'auto'; // Expands collapsed elements el.style.width = 'auto'; // Ensures full visibility el.removeAttribute('hidden'); // Removes hidden attribute el.classList.remove('hidden', 'd-none'); // Removes common CSS hidden classes }) """)
Recursively makes all child elements inside the given selector fully visible.
action_make_all_child_elements_visible
python
dgtlmoon/changedetection.io
changedetectionio/blueprint/browser_steps/browser_steps.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/blueprint/browser_steps/browser_steps.py
Apache-2.0
async def cleanup(self): """Properly clean up all resources to prevent memory leaks""" if self._is_cleaned_up: return logger.debug("Cleaning up browser steps resources") # Clean up page if hasattr(self, 'page') and self.page is not None: try: # Force garbage collection before closing await self.page.request_gc() except Exception as e: logger.debug(f"Error during page garbage collection: {str(e)}") try: # Remove event listeners before closing self.page.remove_listener("close", self.mark_as_closed) except Exception as e: logger.debug(f"Error removing event listeners: {str(e)}") try: await self.page.close() except Exception as e: logger.debug(f"Error closing page: {str(e)}") self.page = None # Clean up context if hasattr(self, 'context') and self.context is not None: try: await self.context.close() except Exception as e: logger.debug(f"Error closing context: {str(e)}") self.context = None self._is_cleaned_up = True logger.debug("Browser steps resources cleanup complete")
Properly clean up all resources to prevent memory leaks
cleanup
python
dgtlmoon/changedetection.io
changedetectionio/blueprint/browser_steps/browser_steps.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/blueprint/browser_steps/browser_steps.py
Apache-2.0
async def get_current_state(self): """Return the screenshot and interactive elements mapping, generally always called after action_()""" import importlib.resources import json # because we for now only run browser steps in playwright mode (not puppeteer mode) from changedetectionio.content_fetchers.playwright import capture_full_page_async # Safety check - don't proceed if resources are cleaned up if self._is_cleaned_up or self.page is None: logger.warning("Attempted to get current state after cleanup") return (None, None) xpath_element_js = importlib.resources.files("changedetectionio.content_fetchers.res").joinpath('xpath_element_scraper.js').read_text() now = time.time() await self.page.wait_for_timeout(1 * 1000) screenshot = None xpath_data = None try: # Get screenshot first screenshot = await capture_full_page_async(page=self.page) if not screenshot: logger.error("No screenshot was retrieved :((") logger.debug(f"Time to get screenshot from browser {time.time() - now:.2f}s") # Then get interactive elements now = time.time() await self.page.evaluate("var include_filters=''") await self.page.request_gc() scan_elements = 'a,button,input,select,textarea,i,th,td,p,li,h1,h2,h3,h4,div,span' MAX_TOTAL_HEIGHT = int(os.getenv("SCREENSHOT_MAX_HEIGHT", SCREENSHOT_MAX_HEIGHT_DEFAULT)) xpath_data = json.loads(await self.page.evaluate(xpath_element_js, { "visualselector_xpath_selectors": scan_elements, "max_height": MAX_TOTAL_HEIGHT })) await self.page.request_gc() # Sort elements by size xpath_data['size_pos'] = sorted(xpath_data['size_pos'], key=lambda k: k['width'] * k['height'], reverse=True) logger.debug(f"Time to scrape xPath element data in browser {time.time()-now:.2f}s") except Exception as e: logger.error(f"Error getting current state: {str(e)}") # If the page has navigated (common with logins) then the context is destroyed on navigation, continue # I'm not sure that this is required anymore because we have the "expect navigation wrapper" at the top if "Execution context was destroyed" in str(e): logger.debug("Execution context was destroyed, most likely because of navigation, continuing...") pass # Attempt recovery - force garbage collection try: await self.page.request_gc() except: pass # Request garbage collection one final time try: await self.page.request_gc() except: pass return (screenshot, xpath_data)
Return the screenshot and interactive elements mapping, generally always called after action_()
get_current_state
python
dgtlmoon/changedetection.io
changedetectionio/blueprint/browser_steps/browser_steps.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/blueprint/browser_steps/browser_steps.py
Apache-2.0
def run_async_in_browser_loop(coro): """Run async coroutine using the existing async worker event loop""" from changedetectionio import worker_handler # Use the existing async worker event loop instead of creating a new one if worker_handler.USE_ASYNC_WORKERS and worker_handler.async_loop and not worker_handler.async_loop.is_closed(): logger.debug("Browser steps using existing async worker event loop") future = asyncio.run_coroutine_threadsafe(coro, worker_handler.async_loop) return future.result() else: # Fallback: create a new event loop (for sync workers or if async loop not available) logger.debug("Browser steps creating temporary event loop") loop = asyncio.new_event_loop() asyncio.set_event_loop(loop) try: return loop.run_until_complete(coro) finally: loop.close()
Run async coroutine using the existing async worker event loop
run_async_in_browser_loop
python
dgtlmoon/changedetection.io
changedetectionio/blueprint/browser_steps/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/blueprint/browser_steps/__init__.py
Apache-2.0
def _watch_has_tag_options_set(watch): """This should be fixed better so that Tag is some proper Model, a tag is just a Watch also""" for tag_uuid, tag in datastore.data['settings']['application'].get('tags', {}).items(): if tag_uuid in watch.get('tags', []) and (tag.get('include_filters') or tag.get('subtractive_selectors')): return True
This should be fixed better so that Tag is some proper Model, a tag is just a Watch also
_watch_has_tag_options_set
python
dgtlmoon/changedetection.io
changedetectionio/blueprint/ui/edit.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/blueprint/ui/edit.py
Apache-2.0
def watch_get_preview_rendered(uuid): '''For when viewing the "preview" of the rendered text from inside of Edit''' from flask import jsonify from changedetectionio.processors.text_json_diff import prepare_filter_prevew result = prepare_filter_prevew(watch_uuid=uuid, form_data=request.form, datastore=datastore) return jsonify(result)
For when viewing the "preview" of the rendered text from inside of Edit
watch_get_preview_rendered
python
dgtlmoon/changedetection.io
changedetectionio/blueprint/ui/edit.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/blueprint/ui/edit.py
Apache-2.0
def form_share_put_watch(uuid): """Given a watch UUID, upload the info and return a share-link the share-link can be imported/added""" import requests import json from copy import deepcopy # more for testing if uuid == 'first': uuid = list(datastore.data['watching'].keys()).pop() # copy it to memory as trim off what we dont need (history) watch = deepcopy(datastore.data['watching'].get(uuid)) # For older versions that are not a @property if (watch.get('history')): del (watch['history']) # for safety/privacy for k in list(watch.keys()): if k.startswith('notification_'): del watch[k] for r in['uuid', 'last_checked', 'last_changed']: if watch.get(r): del (watch[r]) # Add the global stuff which may have an impact watch['ignore_text'] += datastore.data['settings']['application']['global_ignore_text'] watch['subtractive_selectors'] += datastore.data['settings']['application']['global_subtractive_selectors'] watch_json = json.dumps(watch) try: r = requests.request(method="POST", data={'watch': watch_json}, url="https://changedetection.io/share/share", headers={'App-Guid': datastore.data['app_guid']}) res = r.json() # Add to the flask session session['share-link'] = f"https://changedetection.io/share/{res['share_key']}" except Exception as e: logger.error(f"Error sharing -{str(e)}") flash(f"Could not share, something went wrong while communicating with the share server - {str(e)}", 'error') return redirect(url_for('watchlist.index'))
Given a watch UUID, upload the info and return a share-link the share-link can be imported/added
form_share_put_watch
python
dgtlmoon/changedetection.io
changedetectionio/blueprint/ui/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/blueprint/ui/__init__.py
Apache-2.0
def verify_condition_single_rule(watch_uuid): """Verify a single condition rule against the current snapshot""" from changedetectionio.processors.text_json_diff import prepare_filter_prevew from flask import request, jsonify from copy import deepcopy ephemeral_data = {} # Get the watch data watch = datastore.data['watching'].get(watch_uuid) if not watch: return jsonify({'status': 'error', 'message': 'Watch not found'}), 404 # First use prepare_filter_prevew to process the form data # This will return text_after_filter which is after all current form settings are applied # Create ephemeral data with the text from the current snapshot try: # Call prepare_filter_prevew to get a processed version of the content with current form settings # We'll ignore the returned response and just use the datastore which is modified by the function # this should apply all filters etc so then we can run the CONDITIONS against the final output text result = prepare_filter_prevew(datastore=datastore, form_data=request.form, watch_uuid=watch_uuid) ephemeral_data['text'] = result.get('after_filter', '') # Create a temporary watch data structure with this single rule tmp_watch_data = deepcopy(datastore.data['watching'].get(watch_uuid)) # Override the conditions in the temporary watch rule_json = request.args.get("rule") rule = json.loads(rule_json) if rule_json else None # Should be key/value of field, operator, value tmp_watch_data['conditions'] = [rule] tmp_watch_data['conditions_match_logic'] = "ALL" # Single rule, so use ALL # Create a temporary application data structure for the rule check temp_app_data = { 'watching': { watch_uuid: tmp_watch_data } } # Execute the rule against the current snapshot with form data result = execute_ruleset_against_all_plugins( current_watch_uuid=watch_uuid, application_datastruct=temp_app_data, ephemeral_data=ephemeral_data ) return jsonify({ 'status': 'success', 'result': result.get('result'), 'data': result.get('executed_data'), 'message': 'Condition passes' if result else 'Condition does not pass' }) except Exception as e: return jsonify({ 'status': 'error', 'message': f'Error verifying condition: {str(e)}' }), 500
Verify a single condition rule against the current snapshot
verify_condition_single_rule
python
dgtlmoon/changedetection.io
changedetectionio/conditions/blueprint.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/conditions/blueprint.py
Apache-2.0
def convert_to_jsonlogic(logic_operator: str, rule_dict: list): """ Convert a structured rule dict into a JSON Logic rule. :param rule_dict: Dictionary containing conditions. :return: JSON Logic rule as a dictionary. """ json_logic_conditions = [] for condition in rule_dict: operator = condition["operator"] field = condition["field"] value = condition["value"] if not operator or operator == 'None' or not value or not field: raise EmptyConditionRuleRowNotUsable() # Convert value to int/float if possible try: if isinstance(value, str) and "." in value and str != "None": value = float(value) else: value = int(value) except (ValueError, TypeError): pass # Keep as a string if conversion fails # Handle different JSON Logic operators properly if operator == "in": json_logic_conditions.append({"in": [value, {"var": field}]}) # value first elif operator in ("!", "!!", "-"): json_logic_conditions.append({operator: [{"var": field}]}) # Unary operators elif operator in ("min", "max", "cat"): json_logic_conditions.append({operator: value}) # Multi-argument operators else: json_logic_conditions.append({operator: [{"var": field}, value]}) # Standard binary operators return {logic_operator: json_logic_conditions} if len(json_logic_conditions) > 1 else json_logic_conditions[0]
Convert a structured rule dict into a JSON Logic rule. :param rule_dict: Dictionary containing conditions. :return: JSON Logic rule as a dictionary.
convert_to_jsonlogic
python
dgtlmoon/changedetection.io
changedetectionio/conditions/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/conditions/__init__.py
Apache-2.0
def execute_ruleset_against_all_plugins(current_watch_uuid: str, application_datastruct, ephemeral_data={} ): """ Build our data and options by calling our plugins then pass it to jsonlogic and see if the conditions pass :param ruleset: JSON Logic rule dictionary. :param extracted_data: Dictionary containing the facts. <-- maybe the app struct+uuid :return: Dictionary of plugin results. """ from json_logic import jsonLogic EXECUTE_DATA = {} result = True watch = application_datastruct['watching'].get(current_watch_uuid) if watch and watch.get("conditions"): logic_operator = "and" if watch.get("conditions_match_logic", "ALL") == "ALL" else "or" complete_rules = filter_complete_rules(watch['conditions']) if complete_rules: # Give all plugins a chance to update the data dict again (that we will test the conditions against) for plugin in plugin_manager.get_plugins(): try: import concurrent.futures import time with concurrent.futures.ThreadPoolExecutor() as executor: future = executor.submit( plugin.add_data, current_watch_uuid=current_watch_uuid, application_datastruct=application_datastruct, ephemeral_data=ephemeral_data ) logger.debug(f"Trying plugin {plugin}....") # Set a timeout of 10 seconds try: new_execute_data = future.result(timeout=10) if new_execute_data and isinstance(new_execute_data, dict): EXECUTE_DATA.update(new_execute_data) except concurrent.futures.TimeoutError: # The plugin took too long, abort processing for this watch raise Exception(f"Plugin {plugin.__class__.__name__} took more than 10 seconds to run.") except Exception as e: # Log the error but continue with the next plugin import logging logging.error(f"Error executing plugin {plugin.__class__.__name__}: {str(e)}") continue # Create the ruleset ruleset = convert_to_jsonlogic(logic_operator=logic_operator, rule_dict=complete_rules) # Pass the custom operations dictionary to jsonLogic if not jsonLogic(logic=ruleset, data=EXECUTE_DATA, operations=CUSTOM_OPERATIONS): result = False return {'executed_data': EXECUTE_DATA, 'result': result}
Build our data and options by calling our plugins then pass it to jsonlogic and see if the conditions pass :param ruleset: JSON Logic rule dictionary. :param extracted_data: Dictionary containing the facts. <-- maybe the app struct+uuid :return: Dictionary of plugin results.
execute_ruleset_against_all_plugins
python
dgtlmoon/changedetection.io
changedetectionio/conditions/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/conditions/__init__.py
Apache-2.0
def collect_ui_edit_stats_extras(watch): """Collect and combine HTML content from all plugins that implement ui_edit_stats_extras""" extras_content = [] for plugin in plugin_manager.get_plugins(): try: content = plugin.ui_edit_stats_extras(watch=watch) if content: extras_content.append(content) except Exception as e: # Skip plugins that don't implement the hook or have errors pass return "\n".join(extras_content) if extras_content else ""
Collect and combine HTML content from all plugins that implement ui_edit_stats_extras
collect_ui_edit_stats_extras
python
dgtlmoon/changedetection.io
changedetectionio/conditions/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/conditions/__init__.py
Apache-2.0
def ui_edit_stats_extras(watch): """Add Levenshtein stats to the UI using the global plugin system""" """Generate the HTML for Levenshtein stats - shared by both plugin systems""" if len(watch.history.keys()) < 2: return "<p>Not enough history to calculate Levenshtein metrics</p>" try: lev_data = levenshtein_ratio_recent_history(watch) if not lev_data or not isinstance(lev_data, dict): return "<p>Unable to calculate Levenshtein metrics</p>" html = f""" <div class="levenshtein-stats"> <h4>Levenshtein Text Similarity Details</h4> <table class="pure-table"> <tbody> <tr> <td>Raw distance (edits needed)</td> <td>{lev_data['distance']}</td> </tr> <tr> <td>Similarity ratio</td> <td>{lev_data['ratio']:.4f}</td> </tr> <tr> <td>Percent similar</td> <td>{lev_data['percent_similar']}%</td> </tr> </tbody> </table> <p style="font-size: 80%;">Levenshtein metrics compare the last two snapshots, measuring how many character edits are needed to transform one into the other.</p> </div> """ return html except Exception as e: logger.error(f"Error generating Levenshtein UI extras: {str(e)}") return "<p>Error calculating Levenshtein metrics</p>"
Add Levenshtein stats to the UI using the global plugin system
ui_edit_stats_extras
python
dgtlmoon/changedetection.io
changedetectionio/conditions/plugins/levenshtein_plugin.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/conditions/plugins/levenshtein_plugin.py
Apache-2.0
def add_data(current_watch_uuid, application_datastruct, ephemeral_data): """Add word count data for conditions""" result = {} watch = application_datastruct['watching'].get(current_watch_uuid) if watch and 'text' in ephemeral_data: word_count = count_words_in_history(watch, ephemeral_data['text']) result['word_count'] = word_count return result
Add word count data for conditions
add_data
python
dgtlmoon/changedetection.io
changedetectionio/conditions/plugins/wordcount_plugin.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/conditions/plugins/wordcount_plugin.py
Apache-2.0
def _generate_stats_html(watch): """Generate the HTML content for the stats tab""" word_count = count_words_in_history(watch) html = f""" <div class="word-count-stats"> <h4>Content Analysis</h4> <table class="pure-table"> <tbody> <tr> <td>Word count (latest snapshot)</td> <td>{word_count}</td> </tr> </tbody> </table> <p style="font-size: 80%;">Word count is a simple measure of content length, calculated by splitting text on whitespace.</p> </div> """ return html
Generate the HTML content for the stats tab
_generate_stats_html
python
dgtlmoon/changedetection.io
changedetectionio/conditions/plugins/wordcount_plugin.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/conditions/plugins/wordcount_plugin.py
Apache-2.0
def manage_user_agent(headers, current_ua=''): """ Basic setting of user-agent NOTE!!!!!! The service that does the actual Chrome fetching should handle any anti-robot techniques THERE ARE MANY WAYS THAT IT CAN BE DETECTED AS A ROBOT!! This does not take care of - Scraping of 'navigator' (platform, productSub, vendor, oscpu etc etc) browser object (navigator.appVersion) etc - TCP/IP fingerprint JA3 etc - Graphic rendering fingerprinting - Your IP being obviously in a pool of bad actors - Too many requests - Scraping of SCH-UA browser replies (thanks google!!) - Scraping of ServiceWorker, new window calls etc See https://filipvitas.medium.com/how-to-set-user-agent-header-with-puppeteer-js-and-not-fail-28c7a02165da Puppeteer requests https://github.com/dgtlmoon/pyppeteerstealth :param page: :param headers: :return: """ # Ask it what the user agent is, if its obviously ChromeHeadless, switch it to the default ua_in_custom_headers = headers.get('User-Agent') if ua_in_custom_headers: return ua_in_custom_headers if not ua_in_custom_headers and current_ua: current_ua = current_ua.replace('HeadlessChrome', 'Chrome') return current_ua return None
Basic setting of user-agent NOTE!!!!!! The service that does the actual Chrome fetching should handle any anti-robot techniques THERE ARE MANY WAYS THAT IT CAN BE DETECTED AS A ROBOT!! This does not take care of - Scraping of 'navigator' (platform, productSub, vendor, oscpu etc etc) browser object (navigator.appVersion) etc - TCP/IP fingerprint JA3 etc - Graphic rendering fingerprinting - Your IP being obviously in a pool of bad actors - Too many requests - Scraping of SCH-UA browser replies (thanks google!!) - Scraping of ServiceWorker, new window calls etc See https://filipvitas.medium.com/how-to-set-user-agent-header-with-puppeteer-js-and-not-fail-28c7a02165da Puppeteer requests https://github.com/dgtlmoon/pyppeteerstealth :param page: :param headers: :return:
manage_user_agent
python
dgtlmoon/changedetection.io
changedetectionio/content_fetchers/base.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/content_fetchers/base.py
Apache-2.0
def _run_sync(self, url, timeout, request_headers, request_body, request_method, ignore_status_codes=False, current_include_filters=None, is_binary=False, empty_pages_are_a_change=False): """Synchronous version of run - the original requests implementation""" import chardet import requests from requests.exceptions import ProxyError, ConnectionError, RequestException if self.browser_steps_get_valid_steps(): raise BrowserStepsInUnsupportedFetcher(url=url) proxies = {} # Allows override the proxy on a per-request basis # https://requests.readthedocs.io/en/latest/user/advanced/#socks # Should also work with `socks5://user:pass@host:port` type syntax. if self.proxy_override: proxies = {'http': self.proxy_override, 'https': self.proxy_override, 'ftp': self.proxy_override} else: if self.system_http_proxy: proxies['http'] = self.system_http_proxy if self.system_https_proxy: proxies['https'] = self.system_https_proxy session = requests.Session() if strtobool(os.getenv('ALLOW_FILE_URI', 'false')) and url.startswith('file://'): from requests_file import FileAdapter session.mount('file://', FileAdapter()) try: r = session.request(method=request_method, data=request_body.encode('utf-8') if type(request_body) is str else request_body, url=url, headers=request_headers, timeout=timeout, proxies=proxies, verify=False) except Exception as e: msg = str(e) if proxies and 'SOCKSHTTPSConnectionPool' in msg: msg = f"Proxy connection failed? {msg}" raise Exception(msg) from e # If the response did not tell us what encoding format to expect, Then use chardet to override what `requests` thinks. # For example - some sites don't tell us it's utf-8, but return utf-8 content # This seems to not occur when using webdriver/selenium, it seems to detect the text encoding more reliably. # https://github.com/psf/requests/issues/1604 good info about requests encoding detection if not is_binary: # Don't run this for PDF (and requests identified as binary) takes a _long_ time if not r.headers.get('content-type') or not 'charset=' in r.headers.get('content-type'): encoding = chardet.detect(r.content)['encoding'] if encoding: r.encoding = encoding self.headers = r.headers if not r.content or not len(r.content): logger.debug(f"Requests returned empty content for '{url}'") if not empty_pages_are_a_change: raise EmptyReply(url=url, status_code=r.status_code) else: logger.debug(f"URL {url} gave zero byte content reply with Status Code {r.status_code}, but empty_pages_are_a_change = True") # @todo test this # @todo maybe you really want to test zero-byte return pages? if r.status_code != 200 and not ignore_status_codes: # maybe check with content works? raise Non200ErrorCodeReceived(url=url, status_code=r.status_code, page_html=r.text) self.status_code = r.status_code if is_binary: # Binary files just return their checksum until we add something smarter self.content = hashlib.md5(r.content).hexdigest() else: self.content = r.text self.raw_content = r.content
Synchronous version of run - the original requests implementation
_run_sync
python
dgtlmoon/changedetection.io
changedetectionio/content_fetchers/requests.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/content_fetchers/requests.py
Apache-2.0
async def run(self, url, timeout, request_headers, request_body, request_method, ignore_status_codes=False, current_include_filters=None, is_binary=False, empty_pages_are_a_change=False): """Async wrapper that runs the synchronous requests code in a thread pool""" loop = asyncio.get_event_loop() # Run the synchronous _run_sync in a thread pool to avoid blocking the event loop await loop.run_in_executor( None, # Use default ThreadPoolExecutor lambda: self._run_sync( url=url, timeout=timeout, request_headers=request_headers, request_body=request_body, request_method=request_method, ignore_status_codes=ignore_status_codes, current_include_filters=current_include_filters, is_binary=is_binary, empty_pages_are_a_change=empty_pages_are_a_change ) )
Async wrapper that runs the synchronous requests code in a thread pool
run
python
dgtlmoon/changedetection.io
changedetectionio/content_fetchers/requests.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/content_fetchers/requests.py
Apache-2.0
def get_fetch_backend(self): """ Like just using the `fetch_backend` key but there could be some logic :return: """ # Maybe also if is_image etc? # This is because chrome/playwright wont render the PDF in the browser and we will just fetch it and use pdf2html to see the text. if self.is_pdf: return 'html_requests' return self.get('fetch_backend')
Like just using the `fetch_backend` key but there could be some logic :return:
get_fetch_backend
python
dgtlmoon/changedetection.io
changedetectionio/model/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/model/Watch.py
Apache-2.0
def history(self): """History index is just a text file as a list {watch-uuid}/history.txt contains a list like {epoch-time},{filename}\n We read in this list as the history information """ tmp_history = {} # In the case we are only using the watch for processing without history if not self.watch_data_dir: return [] # Read the history file as a dict fname = os.path.join(self.watch_data_dir, "history.txt") if os.path.isfile(fname): logger.debug(f"Reading watch history index for {self.get('uuid')}") with open(fname, "r") as f: for i in f.readlines(): if ',' in i: k, v = i.strip().split(',', 2) # The index history could contain a relative path, so we need to make the fullpath # so that python can read it if not '/' in v and not '\'' in v: v = os.path.join(self.watch_data_dir, v) else: # It's possible that they moved the datadir on older versions # So the snapshot exists but is in a different path snapshot_fname = v.split('/')[-1] proposed_new_path = os.path.join(self.watch_data_dir, snapshot_fname) if not os.path.exists(v) and os.path.exists(proposed_new_path): v = proposed_new_path tmp_history[k] = v if len(tmp_history): self.__newest_history_key = list(tmp_history.keys())[-1] else: self.__newest_history_key = None self.__history_n = len(tmp_history) return tmp_history
History index is just a text file as a list {watch-uuid}/history.txt contains a list like {epoch-time},{filename} We read in this list as the history information
history
python
dgtlmoon/changedetection.io
changedetectionio/model/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/model/Watch.py
Apache-2.0
def get_from_version_based_on_last_viewed(self): """Unfortunately for now timestamp is stored as string key""" keys = list(self.history.keys()) if not keys: return None if len(keys) == 1: return keys[0] last_viewed = int(self.get('last_viewed')) sorted_keys = sorted(keys, key=lambda x: int(x)) sorted_keys.reverse() # When the 'last viewed' timestamp is greater than or equal the newest snapshot, return second newest if last_viewed >= int(sorted_keys[0]): return sorted_keys[1] # When the 'last viewed' timestamp is between snapshots, return the older snapshot for newer, older in list(zip(sorted_keys[0:], sorted_keys[1:])): if last_viewed < int(newer) and last_viewed >= int(older): return older # When the 'last viewed' timestamp is less than the oldest snapshot, return oldest return sorted_keys[-1]
Unfortunately for now timestamp is stored as string key
get_from_version_based_on_last_viewed
python
dgtlmoon/changedetection.io
changedetectionio/model/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/model/Watch.py
Apache-2.0
def get_error_text(self): """Return the text saved from a previous request that resulted in a non-200 error""" fname = os.path.join(self.watch_data_dir, "last-error.txt") if os.path.isfile(fname): with open(fname, 'r') as f: return f.read() return False
Return the text saved from a previous request that resulted in a non-200 error
get_error_text
python
dgtlmoon/changedetection.io
changedetectionio/model/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/model/Watch.py
Apache-2.0
def get_error_snapshot(self): """Return path to the screenshot that resulted in a non-200 error""" fname = os.path.join(self.watch_data_dir, "last-error-screenshot.png") if os.path.isfile(fname): return fname return False
Return path to the screenshot that resulted in a non-200 error
get_error_snapshot
python
dgtlmoon/changedetection.io
changedetectionio/model/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/model/Watch.py
Apache-2.0
def get_browsersteps_available_screenshots(self): "For knowing which screenshots are available to show the user in BrowserSteps UI" available = [] for f in Path(self.watch_data_dir).glob('step_before-*.jpeg'): step_n=re.search(r'step_before-(\d+)', f.name) if step_n: available.append(step_n.group(1)) return available
For knowing which screenshots are available to show the user in BrowserSteps UI
get_browsersteps_available_screenshots
python
dgtlmoon/changedetection.io
changedetectionio/model/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/model/Watch.py
Apache-2.0
def compile_error_texts(self, has_proxies=None): """Compile error texts for this watch. Accepts has_proxies parameter to ensure it works even outside app context""" from flask import url_for from markupsafe import Markup output = [] # Initialize as list since we're using append last_error = self.get('last_error','') try: url_for('settings.settings_page') except Exception as e: has_app_context = False else: has_app_context = True # has app+request context, we can use url_for() if has_app_context: if last_error: if '403' in last_error: if has_proxies: output.append(str(Markup(f"{last_error} - <a href=\"{url_for('settings.settings_page', uuid=self.get('uuid'))}\">Try other proxies/location</a>&nbsp;'"))) else: output.append(str(Markup(f"{last_error} - <a href=\"{url_for('settings.settings_page', uuid=self.get('uuid'))}\">Try adding external proxies/locations</a>&nbsp;'"))) else: output.append(str(Markup(last_error))) if self.get('last_notification_error'): output.append(str(Markup(f"<div class=\"notification-error\"><a href=\"{url_for('settings.notification_logs')}\">{ self.get('last_notification_error') }</a></div>"))) else: # Lo_Fi version if last_error: output.append(str(Markup(last_error))) if self.get('last_notification_error'): output.append(str(Markup(self.get('last_notification_error')))) res = "\n".join(output) return res
Compile error texts for this watch. Accepts has_proxies parameter to ensure it works even outside app context
compile_error_texts
python
dgtlmoon/changedetection.io
changedetectionio/model/Watch.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/model/Watch.py
Apache-2.0
def find_sub_packages(package_name): """ Find all sub-packages within the given package. :param package_name: The name of the base package to scan for sub-packages. :return: A list of sub-package names. """ package = importlib.import_module(package_name) return [name for _, name, is_pkg in pkgutil.iter_modules(package.__path__) if is_pkg]
Find all sub-packages within the given package. :param package_name: The name of the base package to scan for sub-packages. :return: A list of sub-package names.
find_sub_packages
python
dgtlmoon/changedetection.io
changedetectionio/processors/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/processors/__init__.py
Apache-2.0
def find_processors(): """ Find all subclasses of DifferenceDetectionProcessor in the specified package. :param package_name: The name of the package to scan for processor modules. :return: A list of (module, class) tuples. """ package_name = "changedetectionio.processors" # Name of the current package/module processors = [] sub_packages = find_sub_packages(package_name) for sub_package in sub_packages: module_name = f"{package_name}.{sub_package}.processor" try: module = importlib.import_module(module_name) # Iterate through all classes in the module for name, obj in inspect.getmembers(module, inspect.isclass): if issubclass(obj, difference_detection_processor) and obj is not difference_detection_processor: processors.append((module, sub_package)) except (ModuleNotFoundError, ImportError) as e: logger.warning(f"Failed to import module {module_name}: {e} (find_processors())") return processors
Find all subclasses of DifferenceDetectionProcessor in the specified package. :param package_name: The name of the package to scan for processor modules. :return: A list of (module, class) tuples.
find_processors
python
dgtlmoon/changedetection.io
changedetectionio/processors/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/processors/__init__.py
Apache-2.0
def available_processors(): """ Get a list of processors by name and description for the UI elements :return: A list :) """ processor_classes = find_processors() available = [] for package, processor_class in processor_classes: available.append((processor_class, package.name)) return available
Get a list of processors by name and description for the UI elements :return: A list :)
available_processors
python
dgtlmoon/changedetection.io
changedetectionio/processors/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/processors/__init__.py
Apache-2.0
def get_itemprop_availability(html_content) -> Restock: """ Kind of funny/cool way to find price/availability in one many different possibilities. Use 'extruct' to find any possible RDFa/microdata/json-ld data, make a JSON string from the output then search it. """ from jsonpath_ng import parse import re now = time.time() import extruct logger.trace(f"Imported extruct module in {time.time() - now:.3f}s") now = time.time() # Extruct is very slow, I'm wondering if some ML is going to be faster (800ms on my i7), 'rdfa' seems to be the heaviest. syntaxes = ['dublincore', 'json-ld', 'microdata', 'microformat', 'opengraph'] try: data = extruct.extract(html_content, syntaxes=syntaxes) except Exception as e: logger.warning(f"Unable to extract data, document parsing with extruct failed with {type(e).__name__} - {str(e)}") return Restock() logger.trace(f"Extruct basic extract of all metadata done in {time.time() - now:.3f}s") # First phase, dead simple scanning of anything that looks useful value = Restock() if data: logger.debug("Using jsonpath to find price/availability/etc") price_parse = parse('$..(price|Price)') pricecurrency_parse = parse('$..(pricecurrency|currency|priceCurrency )') availability_parse = parse('$..(availability|Availability)') price_result = _deduplicate_prices(price_parse.find(data)) if price_result: # Right now, we just support single product items, maybe we will store the whole actual metadata seperately in teh future and # parse that for the UI? if len(price_result) > 1 and len(price_result) > 1: # See of all prices are different, in the case that one product has many embedded data types with the same price # One might have $121.95 and another 121.95 etc logger.warning(f"More than one price found {price_result}, throwing exception, cant use this plugin.") raise MoreThanOnePriceFound() value['price'] = price_result[0] pricecurrency_result = pricecurrency_parse.find(data) if pricecurrency_result: value['currency'] = pricecurrency_result[0].value availability_result = availability_parse.find(data) if availability_result: value['availability'] = availability_result[0].value if value.get('availability'): value['availability'] = re.sub(r'(?i)^(https|http)://schema.org/', '', value.get('availability').strip(' "\'').lower()) if value.get('availability') else None # Second, go dig OpenGraph which is something that jsonpath_ng cant do because of the tuples and double-dots (:) if not value.get('price') or value.get('availability'): logger.debug("Alternatively digging through OpenGraph properties for restock/price info..") jsonpath_expr = parse('$..properties') for match in jsonpath_expr.find(data): if not value.get('price'): value['price'] = _search_prop_by_value([match.value], "price:amount") if not value.get('availability'): value['availability'] = _search_prop_by_value([match.value], "product:availability") if not value.get('currency'): value['currency'] = _search_prop_by_value([match.value], "price:currency") logger.trace(f"Processed with Extruct in {time.time()-now:.3f}s") return value
Kind of funny/cool way to find price/availability in one many different possibilities. Use 'extruct' to find any possible RDFa/microdata/json-ld data, make a JSON string from the output then search it.
get_itemprop_availability
python
dgtlmoon/changedetection.io
changedetectionio/processors/restock_diff/processor.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/processors/restock_diff/processor.py
Apache-2.0
def prepare_filter_prevew(datastore, watch_uuid, form_data): '''Used by @app.route("/edit/<string:uuid>/preview-rendered", methods=['POST'])''' from changedetectionio import forms, html_tools from changedetectionio.model.Watch import model as watch_model from concurrent.futures import ProcessPoolExecutor from copy import deepcopy from flask import request import brotli import importlib import os import time now = time.time() text_after_filter = '' text_before_filter = '' trigger_line_numbers = [] ignore_line_numbers = [] tmp_watch = deepcopy(datastore.data['watching'].get(watch_uuid)) if tmp_watch and tmp_watch.history and os.path.isdir(tmp_watch.watch_data_dir): # Splice in the temporary stuff from the form form = forms.processor_text_json_diff_form(formdata=form_data if request.method == 'POST' else None, data=form_data ) # Only update vars that came in via the AJAX post p = {k: v for k, v in form.data.items() if k in form_data.keys()} tmp_watch.update(p) blank_watch_no_filters = watch_model() blank_watch_no_filters['url'] = tmp_watch.get('url') latest_filename = next(reversed(tmp_watch.history)) html_fname = os.path.join(tmp_watch.watch_data_dir, f"{latest_filename}.html.br") with open(html_fname, 'rb') as f: decompressed_data = brotli.decompress(f.read()).decode('utf-8') if html_fname.endswith('.br') else f.read().decode('utf-8') # Just like a normal change detection except provide a fake "watch" object and dont call .call_browser() processor_module = importlib.import_module("changedetectionio.processors.text_json_diff.processor") update_handler = processor_module.perform_site_check(datastore=datastore, watch_uuid=tmp_watch.get('uuid') # probably not needed anymore anyway? ) # Use the last loaded HTML as the input update_handler.datastore = datastore update_handler.fetcher.content = str(decompressed_data) # str() because playwright/puppeteer/requests return string update_handler.fetcher.headers['content-type'] = tmp_watch.get('content-type') # Process our watch with filters and the HTML from disk, and also a blank watch with no filters but also with the same HTML from disk # Do this as a parallel process because it could take some time with ProcessPoolExecutor(max_workers=2) as executor: future1 = executor.submit(_task, tmp_watch, update_handler) future2 = executor.submit(_task, blank_watch_no_filters, update_handler) text_after_filter = future1.result() text_before_filter = future2.result() try: trigger_line_numbers = html_tools.strip_ignore_text(content=text_after_filter, wordlist=tmp_watch['trigger_text'], mode='line numbers' ) except Exception as e: text_before_filter = f"Error: {str(e)}" try: text_to_ignore = tmp_watch.get('ignore_text', []) + datastore.data['settings']['application'].get('global_ignore_text', []) ignore_line_numbers = html_tools.strip_ignore_text(content=text_after_filter, wordlist=text_to_ignore, mode='line numbers' ) except Exception as e: text_before_filter = f"Error: {str(e)}" logger.trace(f"Parsed in {time.time() - now:.3f}s") return ({ 'after_filter': text_after_filter, 'before_filter': text_before_filter.decode('utf-8') if isinstance(text_before_filter, bytes) else text_before_filter, 'duration': time.time() - now, 'trigger_line_numbers': trigger_line_numbers, 'ignore_line_numbers': ignore_line_numbers, })
Used by @app.route("/edit/<string:uuid>/preview-rendered", methods=['POST'])
prepare_filter_prevew
python
dgtlmoon/changedetection.io
changedetectionio/processors/text_json_diff/__init__.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/processors/text_json_diff/__init__.py
Apache-2.0
def register_watch_operation_handlers(socketio, datastore): """Register Socket.IO event handlers for watch operations""" @socketio.on('watch_operation') def handle_watch_operation(data): """Handle watch operations like pause, mute, recheck via Socket.IO""" try: op = data.get('op') uuid = data.get('uuid') logger.debug(f"Socket.IO: Received watch operation '{op}' for UUID {uuid}") if not op or not uuid: emit('operation_result', {'success': False, 'error': 'Missing operation or UUID'}) return # Check if watch exists if not datastore.data['watching'].get(uuid): emit('operation_result', {'success': False, 'error': 'Watch not found'}) return watch = datastore.data['watching'][uuid] # Perform the operation if op == 'pause': watch.toggle_pause() logger.info(f"Socket.IO: Toggled pause for watch {uuid}") elif op == 'mute': watch.toggle_mute() logger.info(f"Socket.IO: Toggled mute for watch {uuid}") elif op == 'recheck': # Import here to avoid circular imports from changedetectionio.flask_app import update_q from changedetectionio import queuedWatchMetaData from changedetectionio import worker_handler worker_handler.queue_item_async_safe(update_q, queuedWatchMetaData.PrioritizedItem(priority=1, item={'uuid': uuid})) logger.info(f"Socket.IO: Queued recheck for watch {uuid}") else: emit('operation_result', {'success': False, 'error': f'Unknown operation: {op}'}) return # Send signal to update UI watch_check_update = signal('watch_check_update') if watch_check_update: watch_check_update.send(watch_uuid=uuid) # Send success response to client emit('operation_result', {'success': True, 'operation': op, 'uuid': uuid}) except Exception as e: logger.error(f"Socket.IO error in handle_watch_operation: {str(e)}") emit('operation_result', {'success': False, 'error': str(e)})
Register Socket.IO event handlers for watch operations
register_watch_operation_handlers
python
dgtlmoon/changedetection.io
changedetectionio/realtime/events.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/realtime/events.py
Apache-2.0
def handle_watch_operation(data): """Handle watch operations like pause, mute, recheck via Socket.IO""" try: op = data.get('op') uuid = data.get('uuid') logger.debug(f"Socket.IO: Received watch operation '{op}' for UUID {uuid}") if not op or not uuid: emit('operation_result', {'success': False, 'error': 'Missing operation or UUID'}) return # Check if watch exists if not datastore.data['watching'].get(uuid): emit('operation_result', {'success': False, 'error': 'Watch not found'}) return watch = datastore.data['watching'][uuid] # Perform the operation if op == 'pause': watch.toggle_pause() logger.info(f"Socket.IO: Toggled pause for watch {uuid}") elif op == 'mute': watch.toggle_mute() logger.info(f"Socket.IO: Toggled mute for watch {uuid}") elif op == 'recheck': # Import here to avoid circular imports from changedetectionio.flask_app import update_q from changedetectionio import queuedWatchMetaData from changedetectionio import worker_handler worker_handler.queue_item_async_safe(update_q, queuedWatchMetaData.PrioritizedItem(priority=1, item={'uuid': uuid})) logger.info(f"Socket.IO: Queued recheck for watch {uuid}") else: emit('operation_result', {'success': False, 'error': f'Unknown operation: {op}'}) return # Send signal to update UI watch_check_update = signal('watch_check_update') if watch_check_update: watch_check_update.send(watch_uuid=uuid) # Send success response to client emit('operation_result', {'success': True, 'operation': op, 'uuid': uuid}) except Exception as e: logger.error(f"Socket.IO error in handle_watch_operation: {str(e)}") emit('operation_result', {'success': False, 'error': str(e)})
Handle watch operations like pause, mute, recheck via Socket.IO
handle_watch_operation
python
dgtlmoon/changedetection.io
changedetectionio/realtime/events.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/realtime/events.py
Apache-2.0
def polling_emit_running_or_queued_watches_threaded(self): """Threading version of polling for Windows compatibility""" import time import threading logger.info("Queue update thread started (threading mode)") # Import here to avoid circular imports from changedetectionio.flask_app import app from changedetectionio import worker_handler watch_check_update = signal('watch_check_update') # Track previous state to avoid unnecessary emissions previous_running_uuids = set() # Run until app shutdown - check exit flag more frequently for fast shutdown exit_event = getattr(app.config, 'exit', threading.Event()) while not exit_event.is_set(): try: # Get current running UUIDs from async workers running_uuids = set(worker_handler.get_running_uuids()) # Only send updates for UUIDs that changed state newly_running = running_uuids - previous_running_uuids no_longer_running = previous_running_uuids - running_uuids # Send updates for newly running UUIDs (but exit fast if shutdown requested) for uuid in newly_running: if exit_event.is_set(): break logger.trace(f"Threading polling: UUID {uuid} started processing") with app.app_context(): watch_check_update.send(app_context=app, watch_uuid=uuid) time.sleep(0.01) # Small yield # Send updates for UUIDs that finished processing (but exit fast if shutdown requested) if not exit_event.is_set(): for uuid in no_longer_running: if exit_event.is_set(): break logger.trace(f"Threading polling: UUID {uuid} finished processing") with app.app_context(): watch_check_update.send(app_context=app, watch_uuid=uuid) time.sleep(0.01) # Small yield # Update tracking for next iteration previous_running_uuids = running_uuids # Sleep between polling cycles, but check exit flag every 0.5 seconds for fast shutdown for _ in range(20): # 20 * 0.5 = 10 seconds total if exit_event.is_set(): break time.sleep(0.5) except Exception as e: logger.error(f"Error in threading polling: {str(e)}") # Even during error recovery, check for exit quickly for _ in range(1): # 1 * 0.5 = 0.5 seconds if exit_event.is_set(): break time.sleep(0.5) # Check if we're in pytest environment - if so, be more gentle with logging import sys in_pytest = "pytest" in sys.modules or "PYTEST_CURRENT_TEST" in os.environ if not in_pytest: logger.info("Queue update thread stopped (threading mode)")
Threading version of polling for Windows compatibility
polling_emit_running_or_queued_watches_threaded
python
dgtlmoon/changedetection.io
changedetectionio/realtime/socket_server.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/realtime/socket_server.py
Apache-2.0
def handle_watch_update(socketio, **kwargs): """Handle watch update signal from blinker""" try: watch = kwargs.get('watch') datastore = kwargs.get('datastore') # Emit the watch update to all connected clients from changedetectionio.flask_app import update_q from changedetectionio.flask_app import _jinja2_filter_datetime from changedetectionio import worker_handler # Get list of watches that are currently running running_uuids = worker_handler.get_running_uuids() # Get list of watches in the queue queue_list = [] for q_item in update_q.queue: if hasattr(q_item, 'item') and 'uuid' in q_item.item: queue_list.append(q_item.item['uuid']) # Get the error texts from the watch error_texts = watch.compile_error_texts() # Create a simplified watch data object to send to clients watch_data = { 'checking_now': True if watch.get('uuid') in running_uuids else False, 'fetch_time': watch.get('fetch_time'), 'has_error': True if error_texts else False, 'last_changed': watch.get('last_changed'), 'last_checked': watch.get('last_checked'), 'error_text': error_texts, 'history_n': watch.history_n, 'last_checked_text': _jinja2_filter_datetime(watch), 'last_changed_text': timeago.format(int(watch.last_changed), time.time()) if watch.history_n >= 2 and int(watch.last_changed) > 0 else 'Not yet', 'queued': True if watch.get('uuid') in queue_list else False, 'paused': True if watch.get('paused') else False, 'notification_muted': True if watch.get('notification_muted') else False, 'unviewed': watch.has_unviewed, 'uuid': watch.get('uuid'), 'event_timestamp': time.time() } errored_count = 0 for watch_uuid_iter, watch_iter in datastore.data['watching'].items(): if watch_iter.get('last_error'): errored_count += 1 general_stats = { 'count_errors': errored_count, 'has_unviewed': datastore.has_unviewed } # Debug what's being emitted # logger.debug(f"Emitting 'watch_update' event for {watch.get('uuid')}, data: {watch_data}") # Emit to all clients (no 'broadcast' parameter needed - it's the default behavior) socketio.emit("watch_update", {'watch': watch_data, 'general_stats': general_stats}) # Log after successful emit - use watch_data['uuid'] to avoid variable shadowing issues logger.trace(f"Socket.IO: Emitted update for watch {watch_data['uuid']}, Checking now: {watch_data['checking_now']}") except Exception as e: logger.error(f"Socket.IO error in handle_watch_update: {str(e)}")
Handle watch update signal from blinker
handle_watch_update
python
dgtlmoon/changedetection.io
changedetectionio/realtime/socket_server.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/realtime/socket_server.py
Apache-2.0
def init_socketio(app, datastore): """Initialize SocketIO with the main Flask app""" import platform import sys # Platform-specific async_mode selection for better stability system = platform.system().lower() python_version = sys.version_info # Check for SocketIO mode configuration via environment variable # Default is 'threading' for best cross-platform compatibility socketio_mode = os.getenv('SOCKETIO_MODE', 'threading').lower() if socketio_mode == 'gevent': # Use gevent mode (higher concurrency but platform limitations) try: import gevent async_mode = 'gevent' logger.info(f"SOCKETIO_MODE=gevent: Using {async_mode} mode for Socket.IO") except ImportError: async_mode = 'threading' logger.warning(f"SOCKETIO_MODE=gevent but gevent not available, falling back to {async_mode} mode") elif socketio_mode == 'threading': # Use threading mode (default - best compatibility) async_mode = 'threading' logger.info(f"SOCKETIO_MODE=threading: Using {async_mode} mode for Socket.IO") else: # Invalid mode specified, use default async_mode = 'threading' logger.warning(f"Invalid SOCKETIO_MODE='{socketio_mode}', using default {async_mode} mode for Socket.IO") # Log platform info for debugging logger.info(f"Platform: {system}, Python: {python_version.major}.{python_version.minor}, Socket.IO mode: {async_mode}") # Restrict SocketIO CORS to same origin by default, can be overridden with env var cors_origins = os.environ.get('SOCKETIO_CORS_ORIGINS', None) socketio = SocketIO(app, async_mode=async_mode, cors_allowed_origins=cors_origins, # None means same-origin only logger=strtobool(os.getenv('SOCKETIO_LOGGING', 'False')), engineio_logger=strtobool(os.getenv('SOCKETIO_LOGGING', 'False'))) # Set up event handlers logger.info("Socket.IO: Registering connect event handler") @socketio.on('checkbox-operation') def event_checkbox_operations(data): from changedetectionio.blueprint.ui import _handle_operations from changedetectionio import queuedWatchMetaData from changedetectionio import worker_handler from changedetectionio.flask_app import update_q, watch_check_update logger.trace(f"Got checkbox operations event: {data}") datastore = socketio.datastore _handle_operations( op=data.get('op'), uuids=data.get('uuids'), datastore=datastore, extra_data=data.get('extra_data'), worker_handler=worker_handler, update_q=update_q, queuedWatchMetaData=queuedWatchMetaData, watch_check_update=watch_check_update, emit_flash=False ) @socketio.on('connect') def handle_connect(): """Handle client connection""" # logger.info("Socket.IO: CONNECT HANDLER CALLED - Starting connection process") from flask import request from flask_login import current_user from changedetectionio.flask_app import update_q # Access datastore from socketio datastore = socketio.datastore # logger.info(f"Socket.IO: Current user authenticated: {current_user.is_authenticated if hasattr(current_user, 'is_authenticated') else 'No current_user'}") # Check if authentication is required and user is not authenticated has_password_enabled = datastore.data['settings']['application'].get('password') or os.getenv("SALTED_PASS", False) # logger.info(f"Socket.IO: Password enabled: {has_password_enabled}") if has_password_enabled and not current_user.is_authenticated: logger.warning("Socket.IO: Rejecting unauthenticated connection") return False # Reject the connection # Send the current queue size to the newly connected client try: queue_size = update_q.qsize() socketio.emit("queue_size", { "q_length": queue_size, "event_timestamp": time.time() }, room=request.sid) # Send only to this client logger.debug(f"Socket.IO: Sent initial queue size {queue_size} to new client") except Exception as e: logger.error(f"Socket.IO error sending initial queue size: {str(e)}") logger.info("Socket.IO: Client connected") # logger.info("Socket.IO: Registering disconnect event handler") @socketio.on('disconnect') def handle_disconnect(): """Handle client disconnection""" logger.info("Socket.IO: Client disconnected") # Create a dedicated signal handler that will receive signals and emit them to clients signal_handler = SignalHandler(socketio, datastore) # Register watch operation event handlers from .events import register_watch_operation_handlers register_watch_operation_handlers(socketio, datastore) # Store the datastore reference on the socketio object for later use socketio.datastore = datastore # No stop event needed for threading mode - threads check app.config.exit directly # Add a shutdown method to the socketio object def shutdown(): """Shutdown the SocketIO server fast and aggressively""" try: logger.info("Socket.IO: Fast shutdown initiated...") # For threading mode, give the thread a very short time to exit gracefully if hasattr(socketio, 'polling_emitter_thread'): if socketio.polling_emitter_thread.is_alive(): logger.info("Socket.IO: Waiting 1 second for polling thread to stop...") socketio.polling_emitter_thread.join(timeout=1.0) # Only 1 second timeout if socketio.polling_emitter_thread.is_alive(): logger.info("Socket.IO: Polling thread still running after timeout - continuing with shutdown") else: logger.info("Socket.IO: Polling thread stopped quickly") else: logger.info("Socket.IO: Polling thread already stopped") logger.info("Socket.IO: Fast shutdown complete") except Exception as e: logger.error(f"Socket.IO error during shutdown: {str(e)}") # Attach the shutdown method to the socketio object socketio.shutdown = shutdown logger.info("Socket.IO initialized and attached to main Flask app") logger.info(f"Socket.IO: Registered event handlers: {socketio.handlers if hasattr(socketio, 'handlers') else 'No handlers found'}") return socketio
Initialize SocketIO with the main Flask app
init_socketio
python
dgtlmoon/changedetection.io
changedetectionio/realtime/socket_server.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/realtime/socket_server.py
Apache-2.0
def shutdown(): """Shutdown the SocketIO server fast and aggressively""" try: logger.info("Socket.IO: Fast shutdown initiated...") # For threading mode, give the thread a very short time to exit gracefully if hasattr(socketio, 'polling_emitter_thread'): if socketio.polling_emitter_thread.is_alive(): logger.info("Socket.IO: Waiting 1 second for polling thread to stop...") socketio.polling_emitter_thread.join(timeout=1.0) # Only 1 second timeout if socketio.polling_emitter_thread.is_alive(): logger.info("Socket.IO: Polling thread still running after timeout - continuing with shutdown") else: logger.info("Socket.IO: Polling thread stopped quickly") else: logger.info("Socket.IO: Polling thread already stopped") logger.info("Socket.IO: Fast shutdown complete") except Exception as e: logger.error(f"Socket.IO error during shutdown: {str(e)}")
Shutdown the SocketIO server fast and aggressively
shutdown
python
dgtlmoon/changedetection.io
changedetectionio/realtime/socket_server.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/realtime/socket_server.py
Apache-2.0
def test_conditions_with_text_and_number(client, live_server): """Test that both text and number conditions work together with AND logic.""" set_original_response("50") test_url = url_for('test_endpoint', _external=True) # Add our URL to the import page res = client.post( url_for("imports.import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) # Configure the watch with two conditions connected with AND: # 1. The page filtered text must contain "5" (first digit of value) # 2. The extracted number should be >= 20 and <= 100 res = client.post( url_for("ui.ui_edit.edit_page", uuid="first"), data={ "url": test_url, "fetch_backend": "html_requests", "include_filters": ".number-container", "title": "Number AND Text Condition Test", "conditions_match_logic": "ALL", # ALL = AND logic "conditions-0-operator": "in", "conditions-0-field": "page_filtered_text", "conditions-0-value": "5", "conditions-1-operator": ">=", "conditions-1-field": "extracted_number", "conditions-1-value": "20", "conditions-2-operator": "<=", "conditions-2-field": "extracted_number", "conditions-2-value": "100", # So that 'operations' from pluggy discovery are tested "conditions-3-operator": "length_min", "conditions-3-field": "page_filtered_text", "conditions-3-value": "1", # So that 'operations' from pluggy discovery are tested "conditions-4-operator": "length_max", "conditions-4-field": "page_filtered_text", "conditions-4-value": "100", # So that 'operations' from pluggy discovery are tested "conditions-5-operator": "contains_regex", "conditions-5-field": "page_filtered_text", "conditions-5-value": "\d", }, follow_redirects=True ) assert b"Updated watch." in res.data wait_for_all_checks(client) client.get(url_for("ui.mark_all_viewed"), follow_redirects=True) time.sleep(0.2) wait_for_all_checks(client) # Case 1 set_number_in_range_response("70.5") client.get(url_for("ui.form_watch_checknow"), follow_redirects=True) wait_for_all_checks(client) time.sleep(2) # 75 is > 20 and < 100 and contains "5" res = client.get(url_for("watchlist.index")) assert b'unviewed' in res.data # Case 2: Change with one condition violated # Number out of range (150) but contains '5' client.get(url_for("ui.mark_all_viewed"), follow_redirects=True) time.sleep(0.2) set_number_out_of_range_response("150.5") client.get(url_for("ui.form_watch_checknow"), follow_redirects=True) wait_for_all_checks(client) # Should NOT be marked as having changes since not all conditions are met res = client.get(url_for("watchlist.index")) assert b'unviewed' not in res.data res = client.get(url_for("ui.form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data
Test that both text and number conditions work together with AND logic.
test_conditions_with_text_and_number
python
dgtlmoon/changedetection.io
changedetectionio/tests/test_conditions.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/tests/test_conditions.py
Apache-2.0
def test_render_anchor_tag_content_true(client, live_server, measure_memory_usage): """Testing that the link changes are detected when render_anchor_tag_content setting is set to true""" sleep_time_for_fetch_thread = 3 # Give the endpoint time to spin up time.sleep(1) # set original html text set_original_ignore_response() # Goto the settings page, choose to ignore links (dont select/send "application-render_anchor_tag_content") res = client.post( url_for("settings.settings_page"), data={ "requests-time_between_check-minutes": 180, "application-fetch_backend": "html_requests", }, follow_redirects=True, ) assert b"Settings updated." in res.data # Add our URL to the import page test_url = url_for("test_endpoint", _external=True) res = client.post( url_for("imports.import_page"), data={"urls": test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) # Trigger a check client.get(url_for("ui.form_watch_checknow"), follow_redirects=True) # set a new html text with a modified link set_modified_ignore_response() wait_for_all_checks(client) # Trigger a check client.get(url_for("ui.form_watch_checknow"), follow_redirects=True) wait_for_all_checks(client) # We should not see the rendered anchor tag res = client.get(url_for("ui.ui_views.preview_page", uuid="first")) assert '(/modified_link)' not in res.data.decode() # Goto the settings page, ENABLE render anchor tag res = client.post( url_for("settings.settings_page"), data={ "requests-time_between_check-minutes": 180, "application-render_anchor_tag_content": "true", "application-fetch_backend": "html_requests", }, follow_redirects=True, ) assert b"Settings updated." in res.data # Trigger a check client.get(url_for("ui.form_watch_checknow"), follow_redirects=True) # Give the thread time to pick it up wait_for_all_checks(client) # check that the anchor tag content is rendered res = client.get(url_for("ui.ui_views.preview_page", uuid="first")) assert '(/modified_link)' in res.data.decode() # since the link has changed, and we chose to render anchor tag content, # we should detect a change (new 'unviewed' class) res = client.get(url_for("watchlist.index")) assert b"unviewed" in res.data assert b"/test-endpoint" in res.data # Cleanup everything res = client.get(url_for("ui.form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data
Testing that the link changes are detected when render_anchor_tag_content setting is set to true
test_render_anchor_tag_content_true
python
dgtlmoon/changedetection.io
changedetectionio/tests/test_ignorehyperlinks.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/tests/test_ignorehyperlinks.py
Apache-2.0
def test_import_custom_xlsx(client, live_server, measure_memory_usage): """Test can upload a excel spreadsheet and the watches are created correctly""" dirname = os.path.dirname(__file__) filename = os.path.join(dirname, 'import/spreadsheet.xlsx') with open(filename, 'rb') as f: data= { 'file_mapping': 'custom', 'custom_xlsx[col_0]': '1', 'custom_xlsx[col_1]': '3', 'custom_xlsx[col_2]': '5', 'custom_xlsx[col_3]': '4', 'custom_xlsx[col_type_0]': 'title', 'custom_xlsx[col_type_1]': 'url', 'custom_xlsx[col_type_2]': 'include_filters', 'custom_xlsx[col_type_3]': 'interval_minutes', 'xlsx_file': (io.BytesIO(f.read()), 'spreadsheet.xlsx') } res = client.post( url_for("imports.import_page"), data=data, follow_redirects=True, ) assert b'4 imported from custom .xlsx' in res.data # Because this row was actually just a header with no usable URL, we should get an error assert b'Error processing row number 1' in res.data res = client.get( url_for("watchlist.index") ) assert b'Somesite results ABC' in res.data assert b'City news results' in res.data # Just find one to check over for uuid, watch in live_server.app.config['DATASTORE'].data['watching'].items(): if watch.get('title') == 'Somesite results ABC': filters = watch.get('include_filters') assert filters[0] == '/html[1]/body[1]/div[4]/div[1]/div[1]/div[1]||//*[@id=\'content\']/div[3]/div[1]/div[1]||//*[@id=\'content\']/div[1]' assert watch.get('time_between_check') == {'weeks': 0, 'days': 1, 'hours': 6, 'minutes': 24, 'seconds': 0} res = client.get(url_for("ui.form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data
Test can upload a excel spreadsheet and the watches are created correctly
test_import_custom_xlsx
python
dgtlmoon/changedetection.io
changedetectionio/tests/test_import.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/tests/test_import.py
Apache-2.0
def test_import_watchete_xlsx(client, live_server, measure_memory_usage): """Test can upload a excel spreadsheet and the watches are created correctly""" dirname = os.path.dirname(__file__) filename = os.path.join(dirname, 'import/spreadsheet.xlsx') with open(filename, 'rb') as f: data= { 'file_mapping': 'wachete', 'xlsx_file': (io.BytesIO(f.read()), 'spreadsheet.xlsx') } res = client.post( url_for("imports.import_page"), data=data, follow_redirects=True, ) assert b'4 imported from Wachete .xlsx' in res.data res = client.get( url_for("watchlist.index") ) assert b'Somesite results ABC' in res.data assert b'City news results' in res.data # Just find one to check over for uuid, watch in live_server.app.config['DATASTORE'].data['watching'].items(): if watch.get('title') == 'Somesite results ABC': filters = watch.get('include_filters') assert filters[0] == '/html[1]/body[1]/div[4]/div[1]/div[1]/div[1]||//*[@id=\'content\']/div[3]/div[1]/div[1]||//*[@id=\'content\']/div[1]' assert watch.get('time_between_check') == {'weeks': 0, 'days': 1, 'hours': 6, 'minutes': 24, 'seconds': 0} assert watch.get('fetch_backend') == 'html_requests' # Has inactive 'dynamic wachet' if watch.get('title') == 'JS website': assert watch.get('fetch_backend') == 'html_webdriver' # Has active 'dynamic wachet' if watch.get('title') == 'system default website': assert watch.get('fetch_backend') == 'system' # uses default if blank res = client.get(url_for("ui.form_delete", uuid="all"), follow_redirects=True) assert b'Deleted' in res.data
Test can upload a excel spreadsheet and the watches are created correctly
test_import_watchete_xlsx
python
dgtlmoon/changedetection.io
changedetectionio/tests/test_import.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/tests/test_import.py
Apache-2.0
def test_rss_bad_chars_breaking(client, live_server): """This should absolutely trigger the RSS builder to go into worst state mode - source: prefix means no html conversion (which kinda filters out the bad stuff) - Binary data - Very long so that the saving is performed by Brotli (and decoded back to bytes) Otherwise feedgen should support regular unicode """ with open("test-datastore/endpoint-content.txt", "w") as f: ten_kb_string = "A" * 10_000 f.write(ten_kb_string) test_url = url_for('test_endpoint', _external=True) res = client.post( url_for("imports.import_page"), data={"urls": "source:"+test_url}, follow_redirects=True ) assert b"1 Imported" in res.data wait_for_all_checks(client) # Set the bad content with open("test-datastore/endpoint-content.txt", "w") as f: jpeg_bytes = "\xff\xd8\xff\xe0\x00\x10XXXXXXXX\x00\x01\x02\x00\x00\x01\x00\x01\x00\x00" # JPEG header jpeg_bytes += "A" * 10_000 f.write(jpeg_bytes) res = client.get(url_for("ui.form_watch_checknow"), follow_redirects=True) assert b'Queued 1 watch for rechecking.' in res.data wait_for_all_checks(client) rss_token = extract_rss_token_from_UI(client) uuid = next(iter(live_server.app.config['DATASTORE'].data['watching'])) i=0 from loguru import logger # Because chardet could take a long time while i<=10: logger.debug(f"History was {live_server.app.config['DATASTORE'].data['watching'][uuid].history_n}..") if live_server.app.config['DATASTORE'].data['watching'][uuid].history_n ==2: break i+=1 time.sleep(2) assert live_server.app.config['DATASTORE'].data['watching'][uuid].history_n == 2 # Check RSS feed is still working res = client.get( url_for("rss.feed", uuid=uuid, token=rss_token), follow_redirects=False # Important! leave this off! it should not redirect ) assert res.status_code == 200 #assert live_server.app.config['DATASTORE'].data['watching'][uuid].history_n == 2 #assert live_server.app.config['DATASTORE'].data['watching'][uuid].history_n == 2
This should absolutely trigger the RSS builder to go into worst state mode - source: prefix means no html conversion (which kinda filters out the bad stuff) - Binary data - Very long so that the saving is performed by Brotli (and decoded back to bytes) Otherwise feedgen should support regular unicode
test_rss_bad_chars_breaking
python
dgtlmoon/changedetection.io
changedetectionio/tests/test_rss.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/tests/test_rss.py
Apache-2.0
def wait_for_notification_endpoint_output(): '''Apprise can take a few seconds to fire''' #@todo - could check the apprise object directly instead of looking for this file from os.path import isfile for i in range(1, 20): time.sleep(1) if isfile("test-datastore/notification.txt"): return True return False
Apprise can take a few seconds to fire
wait_for_notification_endpoint_output
python
dgtlmoon/changedetection.io
changedetectionio/tests/util.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/tests/util.py
Apache-2.0
def wait_for_all_checks(client=None): """ Waits until the queue is empty and workers are idle. Much faster than the original with adaptive timing. """ from changedetectionio.flask_app import update_q as global_update_q from changedetectionio import worker_handler logger = logging.getLogger() empty_since = None attempt = 0 max_attempts = 150 # Still reasonable upper bound while attempt < max_attempts: # Start with fast checks, slow down if needed if attempt < 10: time.sleep(0.1) # Very fast initial checks elif attempt < 30: time.sleep(0.3) # Medium speed else: time.sleep(0.8) # Slower for persistent issues q_length = global_update_q.qsize() running_uuids = worker_handler.get_running_uuids() any_workers_busy = len(running_uuids) > 0 if q_length == 0 and not any_workers_busy: if empty_since is None: empty_since = time.time() elif time.time() - empty_since >= 0.15: # Shorter wait break else: empty_since = None attempt += 1 time.sleep(0.3)
Waits until the queue is empty and workers are idle. Much faster than the original with adaptive timing.
wait_for_all_checks
python
dgtlmoon/changedetection.io
changedetectionio/tests/util.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/tests/util.py
Apache-2.0
def test_get_auth(url, expected_auth): """Test authentication extraction with various URL formats.""" parsed_url = apprise_parse_url(url) assert _get_auth(parsed_url) == expected_auth
Test authentication extraction with various URL formats.
test_get_auth
python
dgtlmoon/changedetection.io
changedetectionio/tests/apprise/test_apprise_custom_api_call.py
https://github.com/dgtlmoon/changedetection.io/blob/master/changedetectionio/tests/apprise/test_apprise_custom_api_call.py
Apache-2.0