|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import contextlib |
|
import logging |
|
import sys |
|
from typing import Dict, Iterable, Optional, Set |
|
|
|
from typing_extensions import ContextManager |
|
|
|
from twisted.internet import address, reactor |
|
|
|
import synapse |
|
import synapse.events |
|
from synapse.api.errors import HttpResponseException, RequestSendFailed, SynapseError |
|
from synapse.api.urls import ( |
|
CLIENT_API_PREFIX, |
|
FEDERATION_PREFIX, |
|
LEGACY_MEDIA_PREFIX, |
|
MEDIA_PREFIX, |
|
SERVER_KEY_V2_PREFIX, |
|
) |
|
from synapse.app import _base |
|
from synapse.config._base import ConfigError |
|
from synapse.config.homeserver import HomeServerConfig |
|
from synapse.config.logger import setup_logging |
|
from synapse.config.server import ListenerConfig |
|
from synapse.federation import send_queue |
|
from synapse.federation.transport.server import TransportLayerServer |
|
from synapse.handlers.presence import ( |
|
BasePresenceHandler, |
|
PresenceState, |
|
get_interested_parties, |
|
) |
|
from synapse.http.server import JsonResource, OptionsResource |
|
from synapse.http.servlet import RestServlet, parse_json_object_from_request |
|
from synapse.http.site import SynapseSite |
|
from synapse.logging.context import LoggingContext |
|
from synapse.metrics import METRICS_PREFIX, MetricsResource, RegistryProxy |
|
from synapse.metrics.background_process_metrics import run_as_background_process |
|
from synapse.replication.http import REPLICATION_PREFIX, ReplicationRestResource |
|
from synapse.replication.http.presence import ( |
|
ReplicationBumpPresenceActiveTime, |
|
ReplicationPresenceSetState, |
|
) |
|
from synapse.replication.slave.storage._base import BaseSlavedStore |
|
from synapse.replication.slave.storage.account_data import SlavedAccountDataStore |
|
from synapse.replication.slave.storage.appservice import SlavedApplicationServiceStore |
|
from synapse.replication.slave.storage.client_ips import SlavedClientIpStore |
|
from synapse.replication.slave.storage.deviceinbox import SlavedDeviceInboxStore |
|
from synapse.replication.slave.storage.devices import SlavedDeviceStore |
|
from synapse.replication.slave.storage.directory import DirectoryStore |
|
from synapse.replication.slave.storage.events import SlavedEventStore |
|
from synapse.replication.slave.storage.filtering import SlavedFilteringStore |
|
from synapse.replication.slave.storage.groups import SlavedGroupServerStore |
|
from synapse.replication.slave.storage.keys import SlavedKeyStore |
|
from synapse.replication.slave.storage.presence import SlavedPresenceStore |
|
from synapse.replication.slave.storage.profile import SlavedProfileStore |
|
from synapse.replication.slave.storage.push_rule import SlavedPushRuleStore |
|
from synapse.replication.slave.storage.pushers import SlavedPusherStore |
|
from synapse.replication.slave.storage.receipts import SlavedReceiptsStore |
|
from synapse.replication.slave.storage.registration import SlavedRegistrationStore |
|
from synapse.replication.slave.storage.room import RoomStore |
|
from synapse.replication.slave.storage.transactions import SlavedTransactionStore |
|
from synapse.replication.tcp.client import ReplicationDataHandler |
|
from synapse.replication.tcp.commands import ClearUserSyncsCommand |
|
from synapse.replication.tcp.streams import ( |
|
AccountDataStream, |
|
DeviceListsStream, |
|
GroupServerStream, |
|
PresenceStream, |
|
PushersStream, |
|
PushRulesStream, |
|
ReceiptsStream, |
|
TagAccountDataStream, |
|
ToDeviceStream, |
|
) |
|
from synapse.rest.admin import register_servlets_for_media_repo |
|
from synapse.rest.client.v1 import events |
|
from synapse.rest.client.v1.initial_sync import InitialSyncRestServlet |
|
from synapse.rest.client.v1.login import LoginRestServlet |
|
from synapse.rest.client.v1.profile import ( |
|
ProfileAvatarURLRestServlet, |
|
ProfileDisplaynameRestServlet, |
|
ProfileRestServlet, |
|
) |
|
from synapse.rest.client.v1.push_rule import PushRuleRestServlet |
|
from synapse.rest.client.v1.room import ( |
|
JoinedRoomMemberListRestServlet, |
|
JoinRoomAliasServlet, |
|
PublicRoomListRestServlet, |
|
RoomEventContextServlet, |
|
RoomInitialSyncRestServlet, |
|
RoomMemberListRestServlet, |
|
RoomMembershipRestServlet, |
|
RoomMessageListRestServlet, |
|
RoomSendEventRestServlet, |
|
RoomStateEventRestServlet, |
|
RoomStateRestServlet, |
|
RoomTypingRestServlet, |
|
) |
|
from synapse.rest.client.v1.voip import VoipRestServlet |
|
from synapse.rest.client.v2_alpha import groups, sync, user_directory |
|
from synapse.rest.client.v2_alpha._base import client_patterns |
|
from synapse.rest.client.v2_alpha.account import ThreepidRestServlet |
|
from synapse.rest.client.v2_alpha.account_data import ( |
|
AccountDataServlet, |
|
RoomAccountDataServlet, |
|
) |
|
from synapse.rest.client.v2_alpha.keys import KeyChangesServlet, KeyQueryServlet |
|
from synapse.rest.client.v2_alpha.register import RegisterRestServlet |
|
from synapse.rest.client.versions import VersionsRestServlet |
|
from synapse.rest.health import HealthResource |
|
from synapse.rest.key.v2 import KeyApiV2Resource |
|
from synapse.server import HomeServer, cache_in_self |
|
from synapse.storage.databases.main.censor_events import CensorEventsStore |
|
from synapse.storage.databases.main.client_ips import ClientIpWorkerStore |
|
from synapse.storage.databases.main.media_repository import MediaRepositoryStore |
|
from synapse.storage.databases.main.metrics import ServerMetricsStore |
|
from synapse.storage.databases.main.monthly_active_users import ( |
|
MonthlyActiveUsersWorkerStore, |
|
) |
|
from synapse.storage.databases.main.presence import UserPresenceState |
|
from synapse.storage.databases.main.search import SearchWorkerStore |
|
from synapse.storage.databases.main.stats import StatsStore |
|
from synapse.storage.databases.main.transactions import TransactionWorkerStore |
|
from synapse.storage.databases.main.ui_auth import UIAuthWorkerStore |
|
from synapse.storage.databases.main.user_directory import UserDirectoryStore |
|
from synapse.types import ReadReceipt |
|
from synapse.util.async_helpers import Linearizer |
|
from synapse.util.httpresourcetree import create_resource_tree |
|
from synapse.util.manhole import manhole |
|
from synapse.util.versionstring import get_version_string |
|
|
|
logger = logging.getLogger("synapse.app.generic_worker") |
|
|
|
|
|
class PresenceStatusStubServlet(RestServlet): |
|
"""If presence is disabled this servlet can be used to stub out setting |
|
presence status. |
|
""" |
|
|
|
PATTERNS = client_patterns("/presence/(?P<user_id>[^/]*)/status") |
|
|
|
def __init__(self, hs): |
|
super().__init__() |
|
self.auth = hs.get_auth() |
|
|
|
async def on_GET(self, request, user_id): |
|
await self.auth.get_user_by_req(request) |
|
return 200, {"presence": "offline"} |
|
|
|
async def on_PUT(self, request, user_id): |
|
await self.auth.get_user_by_req(request) |
|
return 200, {} |
|
|
|
|
|
class KeyUploadServlet(RestServlet): |
|
"""An implementation of the `KeyUploadServlet` that responds to read only |
|
requests, but otherwise proxies through to the master instance. |
|
""" |
|
|
|
PATTERNS = client_patterns("/keys/upload(/(?P<device_id>[^/]+))?$") |
|
|
|
def __init__(self, hs): |
|
""" |
|
Args: |
|
hs (synapse.server.HomeServer): server |
|
""" |
|
super().__init__() |
|
self.auth = hs.get_auth() |
|
self.store = hs.get_datastore() |
|
self.http_client = hs.get_simple_http_client() |
|
self.main_uri = hs.config.worker_main_http_uri |
|
|
|
async def on_POST(self, request, device_id): |
|
requester = await self.auth.get_user_by_req(request, allow_guest=True) |
|
user_id = requester.user.to_string() |
|
body = parse_json_object_from_request(request) |
|
|
|
if device_id is not None: |
|
|
|
|
|
if requester.device_id is not None and device_id != requester.device_id: |
|
logger.warning( |
|
"Client uploading keys for a different device " |
|
"(logged in as %s, uploading for %s)", |
|
requester.device_id, |
|
device_id, |
|
) |
|
else: |
|
device_id = requester.device_id |
|
|
|
if device_id is None: |
|
raise SynapseError( |
|
400, "To upload keys, you must pass device_id when authenticating" |
|
) |
|
|
|
if body: |
|
|
|
|
|
|
|
|
|
|
|
headers = { |
|
header: request.requestHeaders.getRawHeaders(header, []) |
|
for header in (b"Authorization", b"User-Agent") |
|
} |
|
|
|
x_forwarded_for = request.requestHeaders.getRawHeaders( |
|
b"X-Forwarded-For", [] |
|
) |
|
if isinstance(request.client, (address.IPv4Address, address.IPv6Address)): |
|
previous_host = request.client.host.encode("ascii") |
|
|
|
|
|
if x_forwarded_for: |
|
x_forwarded_for = [ |
|
x_forwarded_for[0] + b", " + previous_host |
|
] + x_forwarded_for[1:] |
|
else: |
|
x_forwarded_for = [previous_host] |
|
headers[b"X-Forwarded-For"] = x_forwarded_for |
|
|
|
try: |
|
result = await self.http_client.post_json_get_json( |
|
self.main_uri + request.uri.decode("ascii"), body, headers=headers |
|
) |
|
except HttpResponseException as e: |
|
raise e.to_synapse_error() from e |
|
except RequestSendFailed as e: |
|
raise SynapseError(502, "Failed to talk to master") from e |
|
|
|
return 200, result |
|
else: |
|
|
|
result = await self.store.count_e2e_one_time_keys(user_id, device_id) |
|
return 200, {"one_time_key_counts": result} |
|
|
|
|
|
class _NullContextManager(ContextManager[None]): |
|
"""A context manager which does nothing.""" |
|
|
|
def __exit__(self, exc_type, exc_val, exc_tb): |
|
pass |
|
|
|
|
|
UPDATE_SYNCING_USERS_MS = 10 * 1000 |
|
|
|
|
|
class GenericWorkerPresence(BasePresenceHandler): |
|
def __init__(self, hs): |
|
super().__init__(hs) |
|
self.hs = hs |
|
self.is_mine_id = hs.is_mine_id |
|
|
|
|
|
|
|
self._presence_enabled = hs.config.use_presence |
|
|
|
|
|
|
|
self._user_to_num_current_syncs = {} |
|
|
|
self.notifier = hs.get_notifier() |
|
self.instance_id = hs.get_instance_id() |
|
|
|
|
|
|
|
self.users_going_offline = {} |
|
|
|
self._bump_active_client = ReplicationBumpPresenceActiveTime.make_client(hs) |
|
self._set_state_client = ReplicationPresenceSetState.make_client(hs) |
|
|
|
self._send_stop_syncing_loop = self.clock.looping_call( |
|
self.send_stop_syncing, UPDATE_SYNCING_USERS_MS |
|
) |
|
|
|
hs.get_reactor().addSystemEventTrigger( |
|
"before", |
|
"shutdown", |
|
run_as_background_process, |
|
"generic_presence.on_shutdown", |
|
self._on_shutdown, |
|
) |
|
|
|
def _on_shutdown(self): |
|
if self._presence_enabled: |
|
self.hs.get_tcp_replication().send_command( |
|
ClearUserSyncsCommand(self.instance_id) |
|
) |
|
|
|
def send_user_sync(self, user_id, is_syncing, last_sync_ms): |
|
if self._presence_enabled: |
|
self.hs.get_tcp_replication().send_user_sync( |
|
self.instance_id, user_id, is_syncing, last_sync_ms |
|
) |
|
|
|
def mark_as_coming_online(self, user_id): |
|
"""A user has started syncing. Send a UserSync to the master, unless they |
|
had recently stopped syncing. |
|
|
|
Args: |
|
user_id (str) |
|
""" |
|
going_offline = self.users_going_offline.pop(user_id, None) |
|
if not going_offline: |
|
|
|
self.send_user_sync(user_id, True, self.clock.time_msec()) |
|
|
|
def mark_as_going_offline(self, user_id): |
|
"""A user has stopped syncing. We wait before notifying the master as |
|
its likely they'll come back soon. This allows us to avoid sending |
|
a stopped syncing immediately followed by a started syncing notification |
|
to the master |
|
|
|
Args: |
|
user_id (str) |
|
""" |
|
self.users_going_offline[user_id] = self.clock.time_msec() |
|
|
|
def send_stop_syncing(self): |
|
"""Check if there are any users who have stopped syncing a while ago |
|
and haven't come back yet. If there are poke the master about them. |
|
""" |
|
now = self.clock.time_msec() |
|
for user_id, last_sync_ms in list(self.users_going_offline.items()): |
|
if now - last_sync_ms > UPDATE_SYNCING_USERS_MS: |
|
self.users_going_offline.pop(user_id, None) |
|
self.send_user_sync(user_id, False, last_sync_ms) |
|
|
|
async def user_syncing( |
|
self, user_id: str, affect_presence: bool |
|
) -> ContextManager[None]: |
|
"""Record that a user is syncing. |
|
|
|
Called by the sync and events servlets to record that a user has connected to |
|
this worker and is waiting for some events. |
|
""" |
|
if not affect_presence or not self._presence_enabled: |
|
return _NullContextManager() |
|
|
|
curr_sync = self._user_to_num_current_syncs.get(user_id, 0) |
|
self._user_to_num_current_syncs[user_id] = curr_sync + 1 |
|
|
|
|
|
if self._user_to_num_current_syncs[user_id] == 1: |
|
self.mark_as_coming_online(user_id) |
|
|
|
def _end(): |
|
|
|
|
|
|
|
if user_id in self._user_to_num_current_syncs: |
|
self._user_to_num_current_syncs[user_id] -= 1 |
|
|
|
|
|
if self._user_to_num_current_syncs[user_id] == 0: |
|
self.mark_as_going_offline(user_id) |
|
|
|
@contextlib.contextmanager |
|
def _user_syncing(): |
|
try: |
|
yield |
|
finally: |
|
_end() |
|
|
|
return _user_syncing() |
|
|
|
async def notify_from_replication(self, states, stream_id): |
|
parties = await get_interested_parties(self.store, states) |
|
room_ids_to_states, users_to_states = parties |
|
|
|
self.notifier.on_new_event( |
|
"presence_key", |
|
stream_id, |
|
rooms=room_ids_to_states.keys(), |
|
users=users_to_states.keys(), |
|
) |
|
|
|
async def process_replication_rows(self, token, rows): |
|
states = [ |
|
UserPresenceState( |
|
row.user_id, |
|
row.state, |
|
row.last_active_ts, |
|
row.last_federation_update_ts, |
|
row.last_user_sync_ts, |
|
row.status_msg, |
|
row.currently_active, |
|
) |
|
for row in rows |
|
] |
|
|
|
for state in states: |
|
self.user_to_current_state[state.user_id] = state |
|
|
|
stream_id = token |
|
await self.notify_from_replication(states, stream_id) |
|
|
|
def get_currently_syncing_users_for_replication(self) -> Iterable[str]: |
|
return [ |
|
user_id |
|
for user_id, count in self._user_to_num_current_syncs.items() |
|
if count > 0 |
|
] |
|
|
|
async def set_state(self, target_user, state, ignore_status_msg=False): |
|
"""Set the presence state of the user. |
|
""" |
|
presence = state["presence"] |
|
|
|
valid_presence = ( |
|
PresenceState.ONLINE, |
|
PresenceState.UNAVAILABLE, |
|
PresenceState.OFFLINE, |
|
) |
|
if presence not in valid_presence: |
|
raise SynapseError(400, "Invalid presence state") |
|
|
|
user_id = target_user.to_string() |
|
|
|
|
|
if not self.hs.config.use_presence: |
|
return |
|
|
|
|
|
await self._set_state_client( |
|
user_id=user_id, state=state, ignore_status_msg=ignore_status_msg |
|
) |
|
|
|
async def bump_presence_active_time(self, user): |
|
"""We've seen the user do something that indicates they're interacting |
|
with the app. |
|
""" |
|
|
|
if not self.hs.config.use_presence: |
|
return |
|
|
|
|
|
user_id = user.to_string() |
|
await self._bump_active_client(user_id=user_id) |
|
|
|
|
|
class GenericWorkerSlavedStore( |
|
|
|
|
|
UserDirectoryStore, |
|
StatsStore, |
|
UIAuthWorkerStore, |
|
SlavedDeviceInboxStore, |
|
SlavedDeviceStore, |
|
SlavedReceiptsStore, |
|
SlavedPushRuleStore, |
|
SlavedGroupServerStore, |
|
SlavedAccountDataStore, |
|
SlavedPusherStore, |
|
CensorEventsStore, |
|
ClientIpWorkerStore, |
|
SlavedEventStore, |
|
SlavedKeyStore, |
|
RoomStore, |
|
DirectoryStore, |
|
SlavedApplicationServiceStore, |
|
SlavedRegistrationStore, |
|
SlavedTransactionStore, |
|
SlavedProfileStore, |
|
SlavedClientIpStore, |
|
SlavedPresenceStore, |
|
SlavedFilteringStore, |
|
MonthlyActiveUsersWorkerStore, |
|
MediaRepositoryStore, |
|
ServerMetricsStore, |
|
SearchWorkerStore, |
|
TransactionWorkerStore, |
|
BaseSlavedStore, |
|
): |
|
pass |
|
|
|
|
|
class GenericWorkerServer(HomeServer): |
|
DATASTORE_CLASS = GenericWorkerSlavedStore |
|
|
|
def _listen_http(self, listener_config: ListenerConfig): |
|
port = listener_config.port |
|
bind_addresses = listener_config.bind_addresses |
|
|
|
assert listener_config.http_options is not None |
|
|
|
site_tag = listener_config.http_options.tag |
|
if site_tag is None: |
|
site_tag = port |
|
|
|
|
|
resources = {"/health": HealthResource()} |
|
|
|
for res in listener_config.http_options.resources: |
|
for name in res.names: |
|
if name == "metrics": |
|
resources[METRICS_PREFIX] = MetricsResource(RegistryProxy) |
|
elif name == "client": |
|
resource = JsonResource(self, canonical_json=False) |
|
|
|
PublicRoomListRestServlet(self).register(resource) |
|
RoomMemberListRestServlet(self).register(resource) |
|
JoinedRoomMemberListRestServlet(self).register(resource) |
|
RoomStateRestServlet(self).register(resource) |
|
RoomEventContextServlet(self).register(resource) |
|
RoomMessageListRestServlet(self).register(resource) |
|
RegisterRestServlet(self).register(resource) |
|
LoginRestServlet(self).register(resource) |
|
ThreepidRestServlet(self).register(resource) |
|
KeyQueryServlet(self).register(resource) |
|
KeyChangesServlet(self).register(resource) |
|
VoipRestServlet(self).register(resource) |
|
PushRuleRestServlet(self).register(resource) |
|
VersionsRestServlet(self).register(resource) |
|
RoomSendEventRestServlet(self).register(resource) |
|
RoomMembershipRestServlet(self).register(resource) |
|
RoomStateEventRestServlet(self).register(resource) |
|
JoinRoomAliasServlet(self).register(resource) |
|
ProfileAvatarURLRestServlet(self).register(resource) |
|
ProfileDisplaynameRestServlet(self).register(resource) |
|
ProfileRestServlet(self).register(resource) |
|
KeyUploadServlet(self).register(resource) |
|
AccountDataServlet(self).register(resource) |
|
RoomAccountDataServlet(self).register(resource) |
|
RoomTypingRestServlet(self).register(resource) |
|
|
|
sync.register_servlets(self, resource) |
|
events.register_servlets(self, resource) |
|
InitialSyncRestServlet(self).register(resource) |
|
RoomInitialSyncRestServlet(self).register(resource) |
|
|
|
user_directory.register_servlets(self, resource) |
|
|
|
|
|
|
|
if not self.config.use_presence: |
|
PresenceStatusStubServlet(self).register(resource) |
|
|
|
groups.register_servlets(self, resource) |
|
|
|
resources.update({CLIENT_API_PREFIX: resource}) |
|
elif name == "federation": |
|
resources.update({FEDERATION_PREFIX: TransportLayerServer(self)}) |
|
elif name == "media": |
|
if self.config.can_load_media_repo: |
|
media_repo = self.get_media_repository_resource() |
|
|
|
|
|
|
|
admin_resource = JsonResource(self, canonical_json=False) |
|
register_servlets_for_media_repo(self, admin_resource) |
|
|
|
resources.update( |
|
{ |
|
MEDIA_PREFIX: media_repo, |
|
LEGACY_MEDIA_PREFIX: media_repo, |
|
"/_synapse/admin": admin_resource, |
|
} |
|
) |
|
else: |
|
logger.warning( |
|
"A 'media' listener is configured but the media" |
|
" repository is disabled. Ignoring." |
|
) |
|
|
|
if name == "openid" and "federation" not in res.names: |
|
|
|
|
|
|
|
resources.update( |
|
{ |
|
FEDERATION_PREFIX: TransportLayerServer( |
|
self, servlet_groups=["openid"] |
|
) |
|
} |
|
) |
|
|
|
if name in ["keys", "federation"]: |
|
resources[SERVER_KEY_V2_PREFIX] = KeyApiV2Resource(self) |
|
|
|
if name == "replication": |
|
resources[REPLICATION_PREFIX] = ReplicationRestResource(self) |
|
|
|
root_resource = create_resource_tree(resources, OptionsResource()) |
|
|
|
_base.listen_tcp( |
|
bind_addresses, |
|
port, |
|
SynapseSite( |
|
"synapse.access.http.%s" % (site_tag,), |
|
site_tag, |
|
listener_config, |
|
root_resource, |
|
self.version_string, |
|
), |
|
reactor=self.get_reactor(), |
|
) |
|
|
|
logger.info("Synapse worker now listening on port %d", port) |
|
|
|
def start_listening(self, listeners: Iterable[ListenerConfig]): |
|
for listener in listeners: |
|
if listener.type == "http": |
|
self._listen_http(listener) |
|
elif listener.type == "manhole": |
|
_base.listen_tcp( |
|
listener.bind_addresses, |
|
listener.port, |
|
manhole( |
|
username="matrix", password="rabbithole", globals={"hs": self} |
|
), |
|
) |
|
elif listener.type == "metrics": |
|
if not self.get_config().enable_metrics: |
|
logger.warning( |
|
( |
|
"Metrics listener configured, but " |
|
"enable_metrics is not True!" |
|
) |
|
) |
|
else: |
|
_base.listen_metrics(listener.bind_addresses, listener.port) |
|
else: |
|
logger.warning("Unsupported listener type: %s", listener.type) |
|
|
|
self.get_tcp_replication().start_replication(self) |
|
|
|
async def remove_pusher(self, app_id, push_key, user_id): |
|
self.get_tcp_replication().send_remove_pusher(app_id, push_key, user_id) |
|
|
|
@cache_in_self |
|
def get_replication_data_handler(self): |
|
return GenericWorkerReplicationHandler(self) |
|
|
|
@cache_in_self |
|
def get_presence_handler(self): |
|
return GenericWorkerPresence(self) |
|
|
|
|
|
class GenericWorkerReplicationHandler(ReplicationDataHandler): |
|
def __init__(self, hs): |
|
super().__init__(hs) |
|
|
|
self.store = hs.get_datastore() |
|
self.presence_handler = hs.get_presence_handler() |
|
self.notifier = hs.get_notifier() |
|
|
|
self.notify_pushers = hs.config.start_pushers |
|
self.pusher_pool = hs.get_pusherpool() |
|
|
|
self.send_handler = None |
|
if hs.config.send_federation: |
|
self.send_handler = FederationSenderHandler(hs) |
|
|
|
async def on_rdata(self, stream_name, instance_name, token, rows): |
|
await super().on_rdata(stream_name, instance_name, token, rows) |
|
await self._process_and_notify(stream_name, instance_name, token, rows) |
|
|
|
async def _process_and_notify(self, stream_name, instance_name, token, rows): |
|
try: |
|
if self.send_handler: |
|
await self.send_handler.process_replication_rows( |
|
stream_name, token, rows |
|
) |
|
|
|
if stream_name == PushRulesStream.NAME: |
|
self.notifier.on_new_event( |
|
"push_rules_key", token, users=[row.user_id for row in rows] |
|
) |
|
elif stream_name in (AccountDataStream.NAME, TagAccountDataStream.NAME): |
|
self.notifier.on_new_event( |
|
"account_data_key", token, users=[row.user_id for row in rows] |
|
) |
|
elif stream_name == ReceiptsStream.NAME: |
|
self.notifier.on_new_event( |
|
"receipt_key", token, rooms=[row.room_id for row in rows] |
|
) |
|
await self.pusher_pool.on_new_receipts( |
|
token, token, {row.room_id for row in rows} |
|
) |
|
elif stream_name == ToDeviceStream.NAME: |
|
entities = [row.entity for row in rows if row.entity.startswith("@")] |
|
if entities: |
|
self.notifier.on_new_event("to_device_key", token, users=entities) |
|
elif stream_name == DeviceListsStream.NAME: |
|
all_room_ids = set() |
|
for row in rows: |
|
if row.entity.startswith("@"): |
|
room_ids = await self.store.get_rooms_for_user(row.entity) |
|
all_room_ids.update(room_ids) |
|
self.notifier.on_new_event("device_list_key", token, rooms=all_room_ids) |
|
elif stream_name == PresenceStream.NAME: |
|
await self.presence_handler.process_replication_rows(token, rows) |
|
elif stream_name == GroupServerStream.NAME: |
|
self.notifier.on_new_event( |
|
"groups_key", token, users=[row.user_id for row in rows] |
|
) |
|
elif stream_name == PushersStream.NAME: |
|
for row in rows: |
|
if row.deleted: |
|
self.stop_pusher(row.user_id, row.app_id, row.pushkey) |
|
else: |
|
await self.start_pusher(row.user_id, row.app_id, row.pushkey) |
|
except Exception: |
|
logger.exception("Error processing replication") |
|
|
|
async def on_position(self, stream_name: str, instance_name: str, token: int): |
|
await super().on_position(stream_name, instance_name, token) |
|
|
|
await self.on_rdata(stream_name, instance_name, token, []) |
|
|
|
def stop_pusher(self, user_id, app_id, pushkey): |
|
if not self.notify_pushers: |
|
return |
|
|
|
key = "%s:%s" % (app_id, pushkey) |
|
pushers_for_user = self.pusher_pool.pushers.get(user_id, {}) |
|
pusher = pushers_for_user.pop(key, None) |
|
if pusher is None: |
|
return |
|
logger.info("Stopping pusher %r / %r", user_id, key) |
|
pusher.on_stop() |
|
|
|
async def start_pusher(self, user_id, app_id, pushkey): |
|
if not self.notify_pushers: |
|
return |
|
|
|
key = "%s:%s" % (app_id, pushkey) |
|
logger.info("Starting pusher %r / %r", user_id, key) |
|
return await self.pusher_pool.start_pusher_by_id(app_id, pushkey, user_id) |
|
|
|
def on_remote_server_up(self, server: str): |
|
"""Called when get a new REMOTE_SERVER_UP command.""" |
|
|
|
|
|
|
|
if self.send_handler: |
|
self.send_handler.wake_destination(server) |
|
|
|
|
|
class FederationSenderHandler: |
|
"""Processes the fedration replication stream |
|
|
|
This class is only instantiate on the worker responsible for sending outbound |
|
federation transactions. It receives rows from the replication stream and forwards |
|
the appropriate entries to the FederationSender class. |
|
""" |
|
|
|
def __init__(self, hs: GenericWorkerServer): |
|
self.store = hs.get_datastore() |
|
self._is_mine_id = hs.is_mine_id |
|
self.federation_sender = hs.get_federation_sender() |
|
self._hs = hs |
|
|
|
|
|
|
|
self.federation_position = None |
|
|
|
self._fed_position_linearizer = Linearizer(name="_fed_position_linearizer") |
|
|
|
def on_start(self): |
|
|
|
|
|
self.federation_sender.notify_new_events( |
|
self.store.get_room_max_stream_ordering() |
|
) |
|
|
|
def wake_destination(self, server: str): |
|
self.federation_sender.wake_destination(server) |
|
|
|
async def process_replication_rows(self, stream_name, token, rows): |
|
|
|
|
|
if stream_name == "federation": |
|
send_queue.process_rows_for_federation(self.federation_sender, rows) |
|
await self.update_token(token) |
|
|
|
|
|
elif stream_name == ReceiptsStream.NAME: |
|
await self._on_new_receipts(rows) |
|
|
|
|
|
elif stream_name == DeviceListsStream.NAME: |
|
|
|
|
|
|
|
hosts = {row.entity for row in rows if not row.entity.startswith("@")} |
|
for host in hosts: |
|
self.federation_sender.send_device_messages(host) |
|
|
|
elif stream_name == ToDeviceStream.NAME: |
|
|
|
|
|
|
|
hosts = {row.entity for row in rows if not row.entity.startswith("@")} |
|
for host in hosts: |
|
self.federation_sender.send_device_messages(host) |
|
|
|
async def _on_new_receipts(self, rows): |
|
""" |
|
Args: |
|
rows (Iterable[synapse.replication.tcp.streams.ReceiptsStream.ReceiptsStreamRow]): |
|
new receipts to be processed |
|
""" |
|
for receipt in rows: |
|
|
|
if not self._is_mine_id(receipt.user_id): |
|
continue |
|
receipt_info = ReadReceipt( |
|
receipt.room_id, |
|
receipt.receipt_type, |
|
receipt.user_id, |
|
[receipt.event_id], |
|
receipt.data, |
|
) |
|
await self.federation_sender.send_read_receipt(receipt_info) |
|
|
|
async def update_token(self, token): |
|
"""Update the record of where we have processed to in the federation stream. |
|
|
|
Called after we have processed a an update received over replication. Sends |
|
a FEDERATION_ACK back to the master, and stores the token that we have processed |
|
in `federation_stream_position` so that we can restart where we left off. |
|
""" |
|
self.federation_position = token |
|
|
|
|
|
|
|
|
|
|
|
if self._fed_position_linearizer.is_queued(None): |
|
|
|
|
|
return |
|
|
|
run_as_background_process("_save_and_send_ack", self._save_and_send_ack) |
|
|
|
async def _save_and_send_ack(self): |
|
"""Save the current federation position in the database and send an ACK |
|
to master with where we're up to. |
|
""" |
|
try: |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
with (await self._fed_position_linearizer.queue(None)): |
|
|
|
|
|
current_position = self.federation_position |
|
|
|
await self.store.update_federation_out_pos( |
|
"federation", current_position |
|
) |
|
|
|
|
|
|
|
self._hs.get_tcp_replication().send_federation_ack(current_position) |
|
except Exception: |
|
logger.exception("Error updating federation stream position") |
|
|
|
|
|
def start(config_options): |
|
try: |
|
config = HomeServerConfig.load_config("Synapse worker", config_options) |
|
except ConfigError as e: |
|
sys.stderr.write("\n" + str(e) + "\n") |
|
sys.exit(1) |
|
|
|
|
|
assert config.worker_app in ( |
|
"synapse.app.appservice", |
|
"synapse.app.client_reader", |
|
"synapse.app.event_creator", |
|
"synapse.app.federation_reader", |
|
"synapse.app.federation_sender", |
|
"synapse.app.frontend_proxy", |
|
"synapse.app.generic_worker", |
|
"synapse.app.media_repository", |
|
"synapse.app.pusher", |
|
"synapse.app.synchrotron", |
|
"synapse.app.user_dir", |
|
) |
|
|
|
if config.worker_app == "synapse.app.appservice": |
|
if config.appservice.notify_appservices: |
|
sys.stderr.write( |
|
"\nThe appservices must be disabled in the main synapse process" |
|
"\nbefore they can be run in a separate worker." |
|
"\nPlease add ``notify_appservices: false`` to the main config" |
|
"\n" |
|
) |
|
sys.exit(1) |
|
|
|
|
|
config.appservice.notify_appservices = True |
|
else: |
|
|
|
config.appservice.notify_appservices = False |
|
|
|
if config.worker_app == "synapse.app.pusher": |
|
if config.server.start_pushers: |
|
sys.stderr.write( |
|
"\nThe pushers must be disabled in the main synapse process" |
|
"\nbefore they can be run in a separate worker." |
|
"\nPlease add ``start_pushers: false`` to the main config" |
|
"\n" |
|
) |
|
sys.exit(1) |
|
|
|
|
|
config.server.start_pushers = True |
|
else: |
|
|
|
config.server.start_pushers = False |
|
|
|
if config.worker_app == "synapse.app.user_dir": |
|
if config.server.update_user_directory: |
|
sys.stderr.write( |
|
"\nThe update_user_directory must be disabled in the main synapse process" |
|
"\nbefore they can be run in a separate worker." |
|
"\nPlease add ``update_user_directory: false`` to the main config" |
|
"\n" |
|
) |
|
sys.exit(1) |
|
|
|
|
|
config.server.update_user_directory = True |
|
else: |
|
|
|
config.server.update_user_directory = False |
|
|
|
if config.worker_app == "synapse.app.federation_sender": |
|
if config.worker.send_federation: |
|
sys.stderr.write( |
|
"\nThe send_federation must be disabled in the main synapse process" |
|
"\nbefore they can be run in a separate worker." |
|
"\nPlease add ``send_federation: false`` to the main config" |
|
"\n" |
|
) |
|
sys.exit(1) |
|
|
|
|
|
config.worker.send_federation = True |
|
else: |
|
|
|
config.worker.send_federation = False |
|
|
|
synapse.events.USE_FROZEN_DICTS = config.use_frozen_dicts |
|
|
|
hs = GenericWorkerServer( |
|
config.server_name, |
|
config=config, |
|
version_string="Synapse/" + get_version_string(synapse), |
|
) |
|
|
|
setup_logging(hs, config, use_worker_options=True) |
|
|
|
hs.setup() |
|
|
|
|
|
|
|
hs.get_replication_streamer() |
|
|
|
reactor.addSystemEventTrigger( |
|
"before", "startup", _base.start, hs, config.worker_listeners |
|
) |
|
|
|
_base.start_worker_reactor("synapse-generic-worker", config) |
|
|
|
|
|
if __name__ == "__main__": |
|
with LoggingContext("main"): |
|
start(sys.argv[1:]) |
|
|