summary refs log tree commit diff
diff options
context:
space:
mode:
authorMathieu Velten <mathieuv@matrix.org>2022-11-22 18:10:40 +0100
committerMathieu Velten <mathieuv@matrix.org>2022-11-22 18:10:40 +0100
commit53ca28a657b18f273495c951d607e9d36ed10943 (patch)
treedecc2c888f9804ab0d325192cf51f3754f322d9e
parentAdd type ignore unused-awaitable to LoopingCall start call (diff)
downloadsynapse-53ca28a657b18f273495c951d607e9d36ed10943.tar.xz
Reformat
-rw-r--r--synapse/_scripts/update_synapse_database.py2
-rw-r--r--synapse/app/phone_stats_home.py4
-rw-r--r--synapse/appservice/scheduler.py8
-rw-r--r--synapse/federation/federation_server.py6
-rw-r--r--synapse/federation/sender/__init__.py4
-rw-r--r--synapse/federation/sender/per_destination_queue.py2
-rw-r--r--synapse/federation/transport/server/_base.py2
-rw-r--r--synapse/handlers/appservice.py6
-rw-r--r--synapse/handlers/deactivate_account.py2
-rw-r--r--synapse/handlers/device.py2
-rw-r--r--synapse/handlers/devicemessage.py2
-rw-r--r--synapse/handlers/federation.py8
-rw-r--r--synapse/handlers/federation_event.py2
-rw-r--r--synapse/handlers/message.py6
-rw-r--r--synapse/handlers/pagination.py6
-rw-r--r--synapse/handlers/presence.py4
-rw-r--r--synapse/handlers/stats.py2
-rw-r--r--synapse/handlers/typing.py6
-rw-r--r--synapse/handlers/user_directory.py2
-rw-r--r--synapse/http/client.py2
-rw-r--r--synapse/http/connectproxyclient.py4
-rw-r--r--synapse/http/matrixfederationclient.py2
-rw-r--r--synapse/http/server.py2
-rw-r--r--synapse/logging/_remote.py2
-rw-r--r--synapse/logging/context.py4
-rw-r--r--synapse/logging/opentracing.py2
-rw-r--r--synapse/metrics/_gc.py2
-rw-r--r--synapse/push/emailpusher.py2
-rw-r--r--synapse/push/httppusher.py4
-rw-r--r--synapse/push/pusherpool.py4
-rw-r--r--synapse/replication/tcp/client.py2
-rw-r--r--synapse/replication/tcp/handler.py2
-rw-r--r--synapse/replication/tcp/protocol.py2
-rw-r--r--synapse/replication/tcp/redis.py6
-rw-r--r--synapse/replication/tcp/resource.py2
-rw-r--r--synapse/rest/client/room.py2
-rw-r--r--synapse/rest/client/transactions.py2
-rw-r--r--synapse/rest/media/v1/storage_provider.py2
-rw-r--r--synapse/storage/background_updates.py2
-rw-r--r--synapse/storage/controllers/persist_events.py2
-rw-r--r--synapse/storage/databases/main/events_worker.py2
-rw-r--r--synapse/util/__init__.py2
-rw-r--r--synapse/util/async_helpers.py16
-rw-r--r--synapse/util/batching_queue.py2
-rw-r--r--synapse/util/caches/cached_call.py2
-rw-r--r--synapse/util/caches/descriptors.py2
-rw-r--r--synapse/util/distributor.py2
-rw-r--r--synapse/util/patch_inline_callbacks.py2
-rw-r--r--synapse/util/ratelimitutils.py6
-rw-r--r--synapse/util/retryutils.py2
-rw-r--r--tests/http/test_matrixfederationclient.py2
-rw-r--r--tests/test_server.py2
-rw-r--r--tests/unittest.py2
53 files changed, 87 insertions, 87 deletions
diff --git a/synapse/_scripts/update_synapse_database.py b/synapse/_scripts/update_synapse_database.py

index 5ae5f9e297..8bf5ad2b36 100644 --- a/synapse/_scripts/update_synapse_database.py +++ b/synapse/_scripts/update_synapse_database.py
@@ -59,7 +59,7 @@ def run_background_updates(hs: HomeServer) -> None: def run() -> None: # Apply all background updates on the database. - defer.ensureDeferred( # type: ignore[unused-awaitable] + defer.ensureDeferred( # type: ignore[unused-awaitable] run_as_background_process("background_updates", run_background_updates) ) diff --git a/synapse/app/phone_stats_home.py b/synapse/app/phone_stats_home.py
index 4599214514..7f67be03b2 100644 --- a/synapse/app/phone_stats_home.py +++ b/synapse/app/phone_stats_home.py
@@ -189,7 +189,7 @@ def start_phone_stats_home(hs: "HomeServer") -> None: clock.looping_call( hs.get_datastores().main.reap_monthly_active_users, 1000 * 60 * 60 ) - hs.get_datastores().main.reap_monthly_active_users() # type: ignore[unused-awaitable] + hs.get_datastores().main.reap_monthly_active_users() # type: ignore[unused-awaitable] @wrap_as_background_process("generate_monthly_active_users") async def generate_monthly_active_users() -> None: @@ -212,7 +212,7 @@ def start_phone_stats_home(hs: "HomeServer") -> None: max_mau_gauge.set(float(hs.config.server.max_mau_value)) if hs.config.server.limit_usage_by_mau or hs.config.server.mau_stats_only: - generate_monthly_active_users() # type: ignore[unused-awaitable] + generate_monthly_active_users() # type: ignore[unused-awaitable] clock.looping_call(generate_monthly_active_users, 5 * 60 * 1000) # End of monthly active user settings diff --git a/synapse/appservice/scheduler.py b/synapse/appservice/scheduler.py
index 8f9f280ec1..32e80d41f2 100644 --- a/synapse/appservice/scheduler.py +++ b/synapse/appservice/scheduler.py
@@ -199,7 +199,7 @@ class _ServiceQueuer: if service.id in self.requests_in_flight: return - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "as-sender-%s" % (service.id,), self._send_request, service ) @@ -406,10 +406,10 @@ class _TransactionController: if sent: await txn.complete(self.store) else: - run_in_background(self._on_txn_fail, service) # type: ignore[unused-awaitable] + run_in_background(self._on_txn_fail, service) # type: ignore[unused-awaitable] except Exception: logger.exception("Error creating appservice transaction") - run_in_background(self._on_txn_fail, service) # type: ignore[unused-awaitable] + run_in_background(self._on_txn_fail, service) # type: ignore[unused-awaitable] async def on_recovered(self, recoverer: "_Recoverer") -> None: logger.info( @@ -478,7 +478,7 @@ class _Recoverer: def recover(self) -> None: def _retry() -> None: - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "as-recoverer-%s" % (self.service.id,), self.retry ) diff --git a/synapse/federation/federation_server.py b/synapse/federation/federation_server.py
index cadd18d18c..a1be969199 100644 --- a/synapse/federation/federation_server.py +++ b/synapse/federation/federation_server.py
@@ -191,7 +191,7 @@ class FederationServer(FederationBase): ) if lock: logger.info("Handling old staged inbound events in %s", room_id) - self._process_incoming_pdus_in_room_inner( # type: ignore[unused-awaitable] + self._process_incoming_pdus_in_room_inner( # type: ignore[unused-awaitable] room_id, room_version, lock, @@ -268,7 +268,7 @@ class FederationServer(FederationBase): # any old events in the staging area. if not self._started_handling_of_staged_events: self._started_handling_of_staged_events = True - self._handle_old_staged_events() # type: ignore[unused-awaitable] + self._handle_old_staged_events() # type: ignore[unused-awaitable] # Start a periodic check for old staged events. This is to handle # the case where locks time out, e.g. if another process gets killed @@ -1129,7 +1129,7 @@ class FederationServer(FederationBase): _INBOUND_EVENT_HANDLING_LOCK_NAME, pdu.room_id ) if lock: - self._process_incoming_pdus_in_room_inner( # type: ignore[unused-awaitable] + self._process_incoming_pdus_in_room_inner( # type: ignore[unused-awaitable] pdu.room_id, room_version, lock, origin, pdu ) diff --git a/synapse/federation/sender/__init__.py b/synapse/federation/sender/__init__.py
index 294be7018b..b6fdd1aa50 100644 --- a/synapse/federation/sender/__init__.py +++ b/synapse/federation/sender/__init__.py
@@ -187,7 +187,7 @@ class _DestinationWakeupQueue: self.queue[destination] = None if not self.processing: - self._handle() # type: ignore[unused-awaitable] + self._handle() # type: ignore[unused-awaitable] @wrap_as_background_process("_DestinationWakeupQueue.handle") async def _handle(self) -> None: @@ -342,7 +342,7 @@ class FederationSender(AbstractFederationSender): return # fire off a processing loop in the background - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "process_event_queue_for_federation", self._process_event_queue_loop ) diff --git a/synapse/federation/sender/per_destination_queue.py b/synapse/federation/sender/per_destination_queue.py
index 9634b75197..b4202d3e55 100644 --- a/synapse/federation/sender/per_destination_queue.py +++ b/synapse/federation/sender/per_destination_queue.py
@@ -262,7 +262,7 @@ class PerDestinationQueue: logger.debug("TX [%s] Starting transaction loop", self._destination) - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "federation_transaction_transmission_loop", self._transaction_transmission_loop, ) diff --git a/synapse/federation/transport/server/_base.py b/synapse/federation/transport/server/_base.py
index 90376dd4f2..2e0c6085af 100644 --- a/synapse/federation/transport/server/_base.py +++ b/synapse/federation/transport/server/_base.py
@@ -132,7 +132,7 @@ class Authenticator: # alive retry_timings = await self.store.get_destination_retry_timings(origin) if retry_timings and retry_timings.retry_last_ts: - run_in_background(self.reset_retry_timings, origin) # type: ignore[unused-awaitable] + run_in_background(self.reset_retry_timings, origin) # type: ignore[unused-awaitable] return origin diff --git a/synapse/handlers/appservice.py b/synapse/handlers/appservice.py
index 94a9f10a2a..4bb33df7ff 100644 --- a/synapse/handlers/appservice.py +++ b/synapse/handlers/appservice.py
@@ -97,7 +97,7 @@ class ApplicationServicesHandler: # We only start a new background process if necessary rather than # optimistically (to cut down on overhead). - self._notify_interested_services(max_token) # type: ignore[unused-awaitable] + self._notify_interested_services(max_token) # type: ignore[unused-awaitable] @wrap_as_background_process("notify_interested_services") async def _notify_interested_services(self, max_token: RoomStreamToken) -> None: @@ -144,7 +144,7 @@ class ApplicationServicesHandler: except Exception: logger.error("Application Services Failure") - run_as_background_process("as_scheduler", start_scheduler) # type: ignore[unused-awaitable] + run_as_background_process("as_scheduler", start_scheduler) # type: ignore[unused-awaitable] self.started_scheduler = True # Fork off pushes to these services @@ -307,7 +307,7 @@ class ApplicationServicesHandler: # We only start a new background process if necessary rather than # optimistically (to cut down on overhead). - self._notify_interested_services_ephemeral( # type: ignore[unused-awaitable] + self._notify_interested_services_ephemeral( # type: ignore[unused-awaitable] services, stream_key, new_token, users ) diff --git a/synapse/handlers/deactivate_account.py b/synapse/handlers/deactivate_account.py
index 94b743d072..4c3fb063b4 100644 --- a/synapse/handlers/deactivate_account.py +++ b/synapse/handlers/deactivate_account.py
@@ -222,7 +222,7 @@ class DeactivateAccountHandler: pending deactivation, if it isn't already running. """ if not self._user_parter_running: - run_as_background_process("user_parter_loop", self._user_parter_loop) # type: ignore[unused-awaitable] + run_as_background_process("user_parter_loop", self._user_parter_loop) # type: ignore[unused-awaitable] async def _user_parter_loop(self) -> None: """Loop that parts deactivated users from rooms""" diff --git a/synapse/handlers/device.py b/synapse/handlers/device.py
index 2a2821300c..55ac7ef612 100644 --- a/synapse/handlers/device.py +++ b/synapse/handlers/device.py
@@ -558,7 +558,7 @@ class DeviceHandler(DeviceWorkerHandler): # We may need to do some processing asynchronously for local user IDs. if self.hs.is_mine_id(user_id): - self._handle_new_device_update_async() # type: ignore[unused-awaitable] + self._handle_new_device_update_async() # type: ignore[unused-awaitable] async def notify_user_signature_update( self, from_user_id: str, user_ids: List[str] diff --git a/synapse/handlers/devicemessage.py b/synapse/handlers/devicemessage.py
index 165ac9249b..1d5de0f29a 100644 --- a/synapse/handlers/devicemessage.py +++ b/synapse/handlers/devicemessage.py
@@ -198,7 +198,7 @@ class DeviceMessageHandler: await self.store.mark_remote_user_device_cache_as_stale(sender_user_id) # Immediately attempt a resync in the background - run_in_background(self._user_device_resync, user_id=sender_user_id) # type: ignore[unused-awaitable] + run_in_background(self._user_device_resync, user_id=sender_user_id) # type: ignore[unused-awaitable] async def send_device_message( self, diff --git a/synapse/handlers/federation.py b/synapse/handlers/federation.py
index 6a44c5d73d..ffc45473c2 100644 --- a/synapse/handlers/federation.py +++ b/synapse/handlers/federation.py
@@ -174,7 +174,7 @@ class FederationHandler: # any partial-state-resync operations which were in flight when we # were shut down. if not hs.config.worker.worker_app: - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "resume_sync_partial_state_room", self._resume_sync_partial_state_room ) @@ -678,7 +678,7 @@ class FederationHandler: if ret.partial_state: # Kick off the process of asynchronously fetching the state for this # room. - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] desc="sync_partial_state_room", func=self._sync_partial_state_room, initial_destination=origin, @@ -722,7 +722,7 @@ class FederationHandler: # lots of requests for missing prev_events which we do actually # have. Hence we fire off the background task, but don't wait for it. - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "handle_queued_pdus", self._handle_queued_pdus, room_queue ) @@ -1626,7 +1626,7 @@ class FederationHandler: partial_state_rooms = await self.store.get_partial_state_room_resync_info() for room_id, resync_info in partial_state_rooms.items(): - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] desc="sync_partial_state_room", func=self._sync_partial_state_room, initial_destination=resync_info.joined_via, diff --git a/synapse/handlers/federation_event.py b/synapse/handlers/federation_event.py
index 60b6af8e98..c8175d66b2 100644 --- a/synapse/handlers/federation_event.py +++ b/synapse/handlers/federation_event.py
@@ -1408,7 +1408,7 @@ class FederationEventHandler: resync = True if resync: - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "resync_device_due_to_pdu", self._resync_device, event.sender, diff --git a/synapse/handlers/message.py b/synapse/handlers/message.py
index 7299aa7e8d..f8599bab2f 100644 --- a/synapse/handlers/message.py +++ b/synapse/handlers/message.py
@@ -101,7 +101,7 @@ class MessageHandler: self._scheduled_expiry: Optional[IDelayedCall] = None if not hs.config.worker.worker_app: - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "_schedule_next_expiry", self._schedule_next_expiry ) @@ -1918,7 +1918,7 @@ class EventCreationHandler: if event.type == EventTypes.Message: # We don't want to block sending messages on any presence code. This # matters as sometimes presence code can take a while. - run_in_background(self._bump_active_time, requester.user) # type: ignore[unused-awaitable] + run_in_background(self._bump_active_time, requester.user) # type: ignore[unused-awaitable] async def _notify() -> None: try: @@ -1928,7 +1928,7 @@ class EventCreationHandler: except Exception: logger.exception("Error notifying about new room events") - run_in_background(_notify) # type: ignore[unused-awaitable] + run_in_background(_notify) # type: ignore[unused-awaitable] return persisted_events[-1] diff --git a/synapse/handlers/pagination.py b/synapse/handlers/pagination.py
index b5043eedf4..8972f58241 100644 --- a/synapse/handlers/pagination.py +++ b/synapse/handlers/pagination.py
@@ -293,7 +293,7 @@ class PaginationHandler: # We want to purge everything, including local events, and to run the purge in # the background so that it's not blocking any other operation apart from # other purges in the same room. - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "_purge_history", self._purge_history, purge_id, @@ -328,7 +328,7 @@ class PaginationHandler: logger.info("[purge] starting purge_id %s", purge_id) self._purges_by_id[purge_id] = PurgeStatus() - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "purge_history", self._purge_history, purge_id, @@ -769,7 +769,7 @@ class PaginationHandler: self._delete_by_id[delete_id] = DeleteStatus() self._delete_by_room.setdefault(room_id, []).append(delete_id) - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "shutdown_and_purge_room", self._shutdown_and_purge_room, delete_id, diff --git a/synapse/handlers/presence.py b/synapse/handlers/presence.py
index 5957f5be3b..7baa45f495 100644 --- a/synapse/handlers/presence.py +++ b/synapse/handlers/presence.py
@@ -1060,7 +1060,7 @@ class PresenceHandler(BasePresenceHandler): yield finally: if affect_presence: - run_in_background(_end) # type: ignore[unused-awaitable] + run_in_background(_end) # type: ignore[unused-awaitable] return _user_syncing() @@ -1333,7 +1333,7 @@ class PresenceHandler(BasePresenceHandler): finally: self._event_processing = False - run_as_background_process("presence.notify_new_event", _process_presence) # type: ignore[unused-awaitable] + run_as_background_process("presence.notify_new_event", _process_presence) # type: ignore[unused-awaitable] async def _unsafe_process(self) -> None: # Loop round handling deltas until we're up to date diff --git a/synapse/handlers/stats.py b/synapse/handlers/stats.py
index 2679def6ea..9a6faf6270 100644 --- a/synapse/handlers/stats.py +++ b/synapse/handlers/stats.py
@@ -75,7 +75,7 @@ class StatsHandler: finally: self._is_processing = False - run_as_background_process("stats.notify_new_event", process) # type: ignore[unused-awaitable] + run_as_background_process("stats.notify_new_event", process) # type: ignore[unused-awaitable] async def _unsafe_process(self) -> None: # If self.pos is None then means we haven't fetched it from DB diff --git a/synapse/handlers/typing.py b/synapse/handlers/typing.py
index d5b142bf94..b59d309606 100644 --- a/synapse/handlers/typing.py +++ b/synapse/handlers/typing.py
@@ -116,7 +116,7 @@ class FollowerTypingHandler: if self.federation and self.is_mine_id(member.user_id): last_fed_poke = self._member_last_federation_poke.get(member, None) if not last_fed_poke or last_fed_poke + FEDERATION_PING_INTERVAL <= now: - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "typing._push_remote", self._push_remote, member=member, typing=True ) @@ -180,7 +180,7 @@ class FollowerTypingHandler: self._room_typing[row.room_id] = now_typing if self.federation: - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "_send_changes_in_typing_to_remotes", self._send_changes_in_typing_to_remotes, row.room_id, @@ -327,7 +327,7 @@ class TypingWriterHandler(FollowerTypingHandler): def _push_update(self, member: RoomMember, typing: bool) -> None: if self.hs.is_mine_id(member.user_id): # Only send updates for changes to our own users. - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "typing._push_remote", self._push_remote, member, typing ) diff --git a/synapse/handlers/user_directory.py b/synapse/handlers/user_directory.py
index b0228d4c7f..96b692dca9 100644 --- a/synapse/handlers/user_directory.py +++ b/synapse/handlers/user_directory.py
@@ -122,7 +122,7 @@ class UserDirectoryHandler(StateDeltasHandler): self._is_processing = False self._is_processing = True - run_as_background_process("user_directory.notify_new_event", process) # type: ignore[unused-awaitable] + run_as_background_process("user_directory.notify_new_event", process) # type: ignore[unused-awaitable] async def handle_local_profile_change( self, user_id: str, profile: ProfileInfo diff --git a/synapse/http/client.py b/synapse/http/client.py
index d7bc09494b..ba56742f1c 100644 --- a/synapse/http/client.py +++ b/synapse/http/client.py
@@ -452,7 +452,7 @@ class SimpleHttpClient: ) # turn timeouts into RequestTimedOutErrors - request_deferred.addErrback(_timeout_to_request_timed_out_error) # type: ignore[unused-awaitable] + request_deferred.addErrback(_timeout_to_request_timed_out_error) # type: ignore[unused-awaitable] response = await make_deferred_yieldable(request_deferred) diff --git a/synapse/http/connectproxyclient.py b/synapse/http/connectproxyclient.py
index 76c01dc268..466e2f57ea 100644 --- a/synapse/http/connectproxyclient.py +++ b/synapse/http/connectproxyclient.py
@@ -102,7 +102,7 @@ class HTTPConnectProxyEndpoint: d = self._proxy_endpoint.connect(f) # once the tcp socket connects successfully, we need to wait for the # CONNECT to complete. - d.addCallback(lambda conn: f.on_connection) # type: ignore[unused-awaitable] + d.addCallback(lambda conn: f.on_connection) # type: ignore[unused-awaitable] return d @@ -196,7 +196,7 @@ class HTTPConnectProtocol(protocol.Protocol): self.http_setup_client = HTTPConnectSetupClient( self.host, self.port, self.proxy_creds ) - self.http_setup_client.on_connected.addCallback(self.proxyConnected) # type: ignore[unused-awaitable] + self.http_setup_client.on_connected.addCallback(self.proxyConnected) # type: ignore[unused-awaitable] def connectionMade(self) -> None: self.http_setup_client.makeConnection(self.transport) diff --git a/synapse/http/matrixfederationclient.py b/synapse/http/matrixfederationclient.py
index aa28de41ea..bf9342fdfe 100644 --- a/synapse/http/matrixfederationclient.py +++ b/synapse/http/matrixfederationclient.py
@@ -1214,7 +1214,7 @@ class MatrixFederationHttpClient: try: d = read_body_with_max_size(response, output_stream, max_size) - d.addTimeout(self.default_timeout, self.reactor) # type: ignore[unused-awaitable] + d.addTimeout(self.default_timeout, self.reactor) # type: ignore[unused-awaitable] length = await make_deferred_yieldable(d) except BodyExceededMaxSize: msg = "Requested file is too large > %r bytes" % (max_size,) diff --git a/synapse/http/server.py b/synapse/http/server.py
index 4605404891..1124d0d0ce 100644 --- a/synapse/http/server.py +++ b/synapse/http/server.py
@@ -757,7 +757,7 @@ def respond_with_json( if send_cors: set_cors_headers(request) - run_in_background( # type: ignore[unused-awaitable] + run_in_background( # type: ignore[unused-awaitable] _async_write_json_to_request_in_thread, request, encoder, json_object ) return NOT_DONE_YET diff --git a/synapse/logging/_remote.py b/synapse/logging/_remote.py
index 3a07df867c..8edf2083fc 100644 --- a/synapse/logging/_remote.py +++ b/synapse/logging/_remote.py
@@ -193,7 +193,7 @@ class RemoteHandler(logging.Handler): self._connection_waiter = None deferred: Deferred = self._service.whenConnected(failAfterFailures=1) - deferred.addCallbacks(writer, fail) # type: ignore[unused-awaitable] + deferred.addCallbacks(writer, fail) # type: ignore[unused-awaitable] self._connection_waiter = deferred def _handle_pressure(self) -> None: diff --git a/synapse/logging/context.py b/synapse/logging/context.py
index 56972f42a6..27041dd870 100644 --- a/synapse/logging/context.py +++ b/synapse/logging/context.py
@@ -843,7 +843,7 @@ def run_in_background( # type: ignore[misc] # which is supposed to have a single entry and exit point. But # by spawning off another deferred, we are effectively # adding a new exit point.) - res.addBoth(_set_context_cb, ctx) # type: ignore[unused-awaitable] + res.addBoth(_set_context_cb, ctx) # type: ignore[unused-awaitable] return res @@ -870,7 +870,7 @@ def make_deferred_yieldable(deferred: "defer.Deferred[T]") -> "defer.Deferred[T] # ok, we can't be sure that a yield won't block, so let's reset the # logcontext, and add a callback to the deferred to restore it. prev_context = set_current_context(SENTINEL_CONTEXT) - deferred.addBoth(_set_context_cb, prev_context) # type: ignore[unused-awaitable] + deferred.addBoth(_set_context_cb, prev_context) # type: ignore[unused-awaitable] return deferred diff --git a/synapse/logging/opentracing.py b/synapse/logging/opentracing.py
index 180061c99e..96d1c9ae9e 100644 --- a/synapse/logging/opentracing.py +++ b/synapse/logging/opentracing.py
@@ -908,7 +908,7 @@ def _custom_sync_async_decorator( scope.__exit__(None, None, None) return result - result.addCallbacks(call_back, err_back) # type: ignore[unused-awaitable] + result.addCallbacks(call_back, err_back) # type: ignore[unused-awaitable] else: if inspect.isawaitable(result): diff --git a/synapse/metrics/_gc.py b/synapse/metrics/_gc.py
index 296e0a59cb..d3aedcaef1 100644 --- a/synapse/metrics/_gc.py +++ b/synapse/metrics/_gc.py
@@ -129,7 +129,7 @@ def install_gc_manager() -> None: gc_unreachable.labels(i).set(unreachable) gc_task = task.LoopingCall(_maybe_gc) - gc_task.start(0.1) # type: ignore[unused-awaitable] + gc_task.start(0.1) # type: ignore[unused-awaitable] # diff --git a/synapse/push/emailpusher.py b/synapse/push/emailpusher.py
index 877a25071d..8b8615673f 100644 --- a/synapse/push/emailpusher.py +++ b/synapse/push/emailpusher.py
@@ -113,7 +113,7 @@ class EmailPusher(Pusher): if self._is_processing: return - run_as_background_process("emailpush.process", self._process) # type: ignore[unused-awaitable] + run_as_background_process("emailpush.process", self._process) # type: ignore[unused-awaitable] def _pause_processing(self) -> None: """Used by tests to temporarily pause processing of events. diff --git a/synapse/push/httppusher.py b/synapse/push/httppusher.py
index e936baf286..5d65194c42 100644 --- a/synapse/push/httppusher.py +++ b/synapse/push/httppusher.py
@@ -160,7 +160,7 @@ class HttpPusher(Pusher): # We could check the receipts are actually m.read receipts here, # but currently that's the only type of receipt anyway... - run_as_background_process("http_pusher.on_new_receipts", self._update_badge) # type: ignore[unused-awaitable] + run_as_background_process("http_pusher.on_new_receipts", self._update_badge) # type: ignore[unused-awaitable] async def _update_badge(self) -> None: # XXX as per https://github.com/matrix-org/matrix-doc/issues/2627, this seems @@ -189,7 +189,7 @@ class HttpPusher(Pusher): if self._is_processing: return - run_as_background_process("httppush.process", self._process) # type: ignore[unused-awaitable] + run_as_background_process("httppush.process", self._process) # type: ignore[unused-awaitable] async def _process(self) -> None: # we should never get here if we are already processing diff --git a/synapse/push/pusherpool.py b/synapse/push/pusherpool.py
index d9589f76f9..4e1bec1110 100644 --- a/synapse/push/pusherpool.py +++ b/synapse/push/pusherpool.py
@@ -92,7 +92,7 @@ class PusherPool: if not self._should_start_pushers: logger.info("Not starting pushers because they are disabled in the config") return - run_as_background_process("start_pushers", self._start_pushers) # type: ignore[unused-awaitable] + run_as_background_process("start_pushers", self._start_pushers) # type: ignore[unused-awaitable] async def add_or_update_pusher( self, @@ -236,7 +236,7 @@ class PusherPool: # We only start a new background process if necessary rather than # optimistically (to cut down on overhead). - self._on_new_notifications(max_token) # type: ignore[unused-awaitable] + self._on_new_notifications(max_token) # type: ignore[unused-awaitable] @wrap_as_background_process("on_new_notifications") async def _on_new_notifications(self, max_token: RoomStreamToken) -> None: diff --git a/synapse/replication/tcp/client.py b/synapse/replication/tcp/client.py
index 086f05d46d..8f3ce1edd3 100644 --- a/synapse/replication/tcp/client.py +++ b/synapse/replication/tcp/client.py
@@ -451,7 +451,7 @@ class FederationSenderHandler: # no need to queue up another task. return - run_as_background_process("_save_and_send_ack", self._save_and_send_ack) # type: ignore[unused-awaitable] + run_as_background_process("_save_and_send_ack", self._save_and_send_ack) # type: ignore[unused-awaitable] async def _save_and_send_ack(self) -> None: """Save the current federation position in the database and send an ACK diff --git a/synapse/replication/tcp/handler.py b/synapse/replication/tcp/handler.py
index 9cc036dd21..ca99381648 100644 --- a/synapse/replication/tcp/handler.py +++ b/synapse/replication/tcp/handler.py
@@ -292,7 +292,7 @@ class ReplicationCommandHandler: return # fire off a background process to start processing the queue. - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "process-replication-data", self._unsafe_process_queue, stream_name ) diff --git a/synapse/replication/tcp/protocol.py b/synapse/replication/tcp/protocol.py
index f08e0cac47..4e39e1c43f 100644 --- a/synapse/replication/tcp/protocol.py +++ b/synapse/replication/tcp/protocol.py
@@ -300,7 +300,7 @@ class BaseReplicationStreamProtocol(LineOnlyReceiver): # if so. if isawaitable(res): - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "replication-" + cmd.get_logcontext_id(), lambda: res ) diff --git a/synapse/replication/tcp/redis.py b/synapse/replication/tcp/redis.py
index 450a14ff17..c5b0bb4e1f 100644 --- a/synapse/replication/tcp/redis.py +++ b/synapse/replication/tcp/redis.py
@@ -112,7 +112,7 @@ class RedisSubscriber(txredisapi.SubscriberProtocol): def connectionMade(self) -> None: logger.info("Connected to redis") super().connectionMade() - run_as_background_process("subscribe-replication", self._send_subscribe) # type: ignore[unused-awaitable] + run_as_background_process("subscribe-replication", self._send_subscribe) # type: ignore[unused-awaitable] async def _send_subscribe(self) -> None: # it's important to make sure that we only send the REPLICATE command once we @@ -183,7 +183,7 @@ class RedisSubscriber(txredisapi.SubscriberProtocol): # if so. if isawaitable(res): - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "replication-" + cmd.get_logcontext_id(), lambda: res ) @@ -205,7 +205,7 @@ class RedisSubscriber(txredisapi.SubscriberProtocol): Args: cmd: The command to send """ - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "send-cmd", self._async_send_command, cmd, bg_start_span=False ) diff --git a/synapse/replication/tcp/resource.py b/synapse/replication/tcp/resource.py
index 53aebf9d46..9eb3e34695 100644 --- a/synapse/replication/tcp/resource.py +++ b/synapse/replication/tcp/resource.py
@@ -137,7 +137,7 @@ class ReplicationStreamer: logger.debug("Notifier poke loop already running") return - run_as_background_process("replication_notifier", self._run_notifier_loop) # type: ignore[unused-awaitable] + run_as_background_process("replication_notifier", self._run_notifier_loop) # type: ignore[unused-awaitable] async def _run_notifier_loop(self) -> None: self.is_looping = True diff --git a/synapse/rest/client/room.py b/synapse/rest/client/room.py
index 3e9d537ece..5f56eb4c3b 100644 --- a/synapse/rest/client/room.py +++ b/synapse/rest/client/room.py
@@ -1079,7 +1079,7 @@ class RoomRedactEventRestServlet(TransactionRestServlet): ) if with_relations: - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "redact_related_events", self._relation_handler.redact_events_related_to, requester=requester, diff --git a/synapse/rest/client/transactions.py b/synapse/rest/client/transactions.py
index dbb1f22f19..48f3f144ea 100644 --- a/synapse/rest/client/transactions.py +++ b/synapse/rest/client/transactions.py
@@ -120,7 +120,7 @@ class HttpTransactionCache: # we deliberately do not propagate the error any further, as we # expect the observers to have reported it. - deferred.addErrback(remove_from_map) # type: ignore[unused-awaitable] + deferred.addErrback(remove_from_map) # type: ignore[unused-awaitable] return make_deferred_yieldable(observable.observe()) diff --git a/synapse/rest/media/v1/storage_provider.py b/synapse/rest/media/v1/storage_provider.py
index 4cbec73884..3f48439af1 100644 --- a/synapse/rest/media/v1/storage_provider.py +++ b/synapse/rest/media/v1/storage_provider.py
@@ -112,7 +112,7 @@ class StorageProviderWrapper(StorageProvider): except Exception: logger.exception("Error storing file") - run_in_background(store) # type: ignore[unused-awaitable] + run_in_background(store) # type: ignore[unused-awaitable] async def fetch(self, path: str, file_info: FileInfo) -> Optional[Responder]: if file_info.url_cache: diff --git a/synapse/storage/background_updates.py b/synapse/storage/background_updates.py
index 9726963090..52e2e35d06 100644 --- a/synapse/storage/background_updates.py +++ b/synapse/storage/background_updates.py
@@ -272,7 +272,7 @@ class BackgroundUpdater: # if we start a new background update, not all updates are done. self._all_done = False sleep = self.sleep_enabled - run_as_background_process( # type: ignore[unused-awaitable] + run_as_background_process( # type: ignore[unused-awaitable] "background_updates", self.run_background_updates, sleep ) diff --git a/synapse/storage/controllers/persist_events.py b/synapse/storage/controllers/persist_events.py
index bf7c264d4d..c737116224 100644 --- a/synapse/storage/controllers/persist_events.py +++ b/synapse/storage/controllers/persist_events.py
@@ -293,7 +293,7 @@ class _EventPeristenceQueue(Generic[_PersistResult]): self._currently_persisting_rooms.discard(room_id) # set handle_queue_loop off in the background - run_as_background_process("persist_events", handle_queue_loop) # type: ignore[unused-awaitable] + run_as_background_process("persist_events", handle_queue_loop) # type: ignore[unused-awaitable] def _get_drainining_queue( self, room_id: str diff --git a/synapse/storage/databases/main/events_worker.py b/synapse/storage/databases/main/events_worker.py
index af5cb94239..7405db5b3d 100644 --- a/synapse/storage/databases/main/events_worker.py +++ b/synapse/storage/databases/main/events_worker.py
@@ -942,7 +942,7 @@ class EventsWorkerStore(SQLBaseStore): should_start = False if should_start: - run_as_background_process("fetch_events", self._fetch_thread) # type: ignore[unused-awaitable] + run_as_background_process("fetch_events", self._fetch_thread) # type: ignore[unused-awaitable] async def _fetch_thread(self) -> None: """Services requests for events from `_event_fetch_list`.""" diff --git a/synapse/util/__init__.py b/synapse/util/__init__.py
index c04bbbca44..714be27d86 100644 --- a/synapse/util/__init__.py +++ b/synapse/util/__init__.py
@@ -132,7 +132,7 @@ class Clock: call = task.LoopingCall(f, *args, **kwargs) call.clock = self._reactor d = call.start(msec / 1000.0, now=False) - d.addErrback(log_failure, "Looping call died", consumeErrors=False) # type: ignore[unused-awaitable] + d.addErrback(log_failure, "Looping call died", consumeErrors=False) # type: ignore[unused-awaitable] return call def call_later( diff --git a/synapse/util/async_helpers.py b/synapse/util/async_helpers.py
index d0ae193624..8a63a73bca 100644 --- a/synapse/util/async_helpers.py +++ b/synapse/util/async_helpers.py
@@ -154,7 +154,7 @@ class ObservableDeferred(Generic[_T], AbstractObservableDeferred[_T]): else: return f - deferred.addCallbacks(callback, errback) # type: ignore[unused-awaitable] + deferred.addCallbacks(callback, errback) # type: ignore[unused-awaitable] def observe(self) -> "defer.Deferred[_T]": """Observe the underlying deferred. @@ -588,7 +588,7 @@ class ReadWriteLock: # writer waiting for us and it completed entirely within the # `new_defer.callback()` call above. if self.key_to_current_writer.get(key) == new_defer: - self.key_to_current_writer.pop(key) # type: ignore[unused-awaitable] + self.key_to_current_writer.pop(key) # type: ignore[unused-awaitable] return _ctx_manager() @@ -646,7 +646,7 @@ def timeout_deferred( raise defer.TimeoutError("Timed out after %gs" % (timeout,)) return value - deferred.addErrback(convert_cancelled) # type: ignore[unused-awaitable] + deferred.addErrback(convert_cancelled) # type: ignore[unused-awaitable] def cancel_timeout(result: _T) -> _T: # stop the pending call to cancel the deferred if it's been fired @@ -654,7 +654,7 @@ def timeout_deferred( delayed_call.cancel() return result - deferred.addBoth(cancel_timeout) # type: ignore[unused-awaitable] + deferred.addBoth(cancel_timeout) # type: ignore[unused-awaitable] def success_cb(val: _T) -> None: if not new_d.called: @@ -664,7 +664,7 @@ def timeout_deferred( if not new_d.called: new_d.errback(val) - deferred.addCallbacks(success_cb, failure_cb) # type: ignore[unused-awaitable] + deferred.addCallbacks(success_cb, failure_cb) # type: ignore[unused-awaitable] return new_d @@ -712,7 +712,7 @@ def stop_cancellation(deferred: "defer.Deferred[T]") -> "defer.Deferred[T]": wrapped with `make_deferred_yieldable`. """ new_deferred: "defer.Deferred[T]" = defer.Deferred() - deferred.chainDeferred(new_deferred) # type: ignore[unused-awaitable] + deferred.chainDeferred(new_deferred) # type: ignore[unused-awaitable] return new_deferred @@ -774,10 +774,10 @@ def delay_cancellation(awaitable: Awaitable[T]) -> Awaitable[T]: new_deferred.pause() new_deferred.errback(Failure(CancelledError())) - deferred.addBoth(lambda _: new_deferred.unpause()) # type: ignore[unused-awaitable] + deferred.addBoth(lambda _: new_deferred.unpause()) # type: ignore[unused-awaitable] new_deferred: "defer.Deferred[T]" = defer.Deferred(handle_cancel) - deferred.chainDeferred(new_deferred) # type: ignore[unused-awaitable] + deferred.chainDeferred(new_deferred) # type: ignore[unused-awaitable] return new_deferred diff --git a/synapse/util/batching_queue.py b/synapse/util/batching_queue.py
index 8b47b1606a..72a13cd1a4 100644 --- a/synapse/util/batching_queue.py +++ b/synapse/util/batching_queue.py
@@ -128,7 +128,7 @@ class BatchingQueue(Generic[V, R]): # If we're not currently processing the key fire off a background # process to start processing. if key not in self._processing_keys: - run_as_background_process(self._name, self._process_queue, key) # type: ignore[unused-awaitable] + run_as_background_process(self._name, self._process_queue, key) # type: ignore[unused-awaitable] with self._number_in_flight_metric.track_inprogress(): return await make_deferred_yieldable(d) diff --git a/synapse/util/caches/cached_call.py b/synapse/util/caches/cached_call.py
index 74b2bc4ebc..4061db56a8 100644 --- a/synapse/util/caches/cached_call.py +++ b/synapse/util/caches/cached_call.py
@@ -89,7 +89,7 @@ class CachedCall(Generic[TV]): def got_result(r: Union[TV, Failure]) -> None: self._result = r - self._deferred.addBoth(got_result) # type: ignore[unused-awaitable] + self._deferred.addBoth(got_result) # type: ignore[unused-awaitable] # TODO: consider cancellation semantics. Currently, if the call to get() # is cancelled, the underlying call will continue (and any future calls diff --git a/synapse/util/caches/descriptors.py b/synapse/util/caches/descriptors.py
index 11707f4317..4264730e4f 100644 --- a/synapse/util/caches/descriptors.py +++ b/synapse/util/caches/descriptors.py
@@ -377,7 +377,7 @@ class DeferredCacheListDescriptor(_CacheDescriptorBase): for k, v in r.items(): results[cache_key_to_arg(k)] = v - pending_deferred.addCallback(update_results) # type: ignore[unused-awaitable] + pending_deferred.addCallback(update_results) # type: ignore[unused-awaitable] cached_defers.append(pending_deferred) if missing: diff --git a/synapse/util/distributor.py b/synapse/util/distributor.py
index d3fcfd59b6..c5019a1074 100644 --- a/synapse/util/distributor.py +++ b/synapse/util/distributor.py
@@ -84,7 +84,7 @@ class Distributor: if name not in self.signals: raise KeyError("%r does not have a signal named %s" % (self, name)) - run_as_background_process(name, self.signals[name].fire, *args, **kwargs) # type: ignore[unused-awaitable] + run_as_background_process(name, self.signals[name].fire, *args, **kwargs) # type: ignore[unused-awaitable] P = ParamSpec("P") diff --git a/synapse/util/patch_inline_callbacks.py b/synapse/util/patch_inline_callbacks.py
index b09eb242c2..c568e25be5 100644 --- a/synapse/util/patch_inline_callbacks.py +++ b/synapse/util/patch_inline_callbacks.py
@@ -108,7 +108,7 @@ def do_patch() -> None: raise Exception(err) return r - res.addBoth(check_ctx) # type: ignore[unused-awaitable] + res.addBoth(check_ctx) # type: ignore[unused-awaitable] return res return wrapped diff --git a/synapse/util/ratelimitutils.py b/synapse/util/ratelimitutils.py
index b653b54492..84337b0796 100644 --- a/synapse/util/ratelimitutils.py +++ b/synapse/util/ratelimitutils.py
@@ -334,7 +334,7 @@ class _PerHostRatelimiter: queue_defer = queue_request() return queue_defer - ret_defer.addBoth(on_wait_finished) # type: ignore[unused-awaitable] + ret_defer.addBoth(on_wait_finished) # type: ignore[unused-awaitable] else: ret_defer = queue_request() @@ -358,8 +358,8 @@ class _PerHostRatelimiter: self.ready_request_queue.pop(request_id, None) return r - ret_defer.addCallbacks(on_start, on_err) # type: ignore[unused-awaitable] - ret_defer.addBoth(on_both) # type: ignore[unused-awaitable] + ret_defer.addCallbacks(on_start, on_err) # type: ignore[unused-awaitable] + ret_defer.addBoth(on_both) # type: ignore[unused-awaitable] return make_deferred_yieldable(ret_defer) def _on_exit(self, request_id: object) -> None: diff --git a/synapse/util/retryutils.py b/synapse/util/retryutils.py
index f42b079c82..9ddc00fddd 100644 --- a/synapse/util/retryutils.py +++ b/synapse/util/retryutils.py
@@ -265,4 +265,4 @@ class RetryDestinationLimiter: logger.exception("Failed to store destination_retry_timings") # we deliberately do this in the background. - run_as_background_process("store_retry_timings", store_retry_timings) # type: ignore[unused-awaitable] + run_as_background_process("store_retry_timings", store_retry_timings) # type: ignore[unused-awaitable] diff --git a/tests/http/test_matrixfederationclient.py b/tests/http/test_matrixfederationclient.py
index 56fb34a1c3..9b64890bdb 100644 --- a/tests/http/test_matrixfederationclient.py +++ b/tests/http/test_matrixfederationclient.py
@@ -451,7 +451,7 @@ class FederationClientTests(HomeserverTestCase): self.failureResultOf(d) def test_client_sends_body(self): - defer.ensureDeferred( # type: ignore[unused-awaitable] + defer.ensureDeferred( # type: ignore[unused-awaitable] self.cl.post_json( "testserv:8008", "foo/bar", timeout=10000, data={"a": "b"} ) diff --git a/tests/test_server.py b/tests/test_server.py
index be85639067..6f35966d0c 100644 --- a/tests/test_server.py +++ b/tests/test_server.py
@@ -117,7 +117,7 @@ class JsonResourceTests(unittest.TestCase): def _callback(request: SynapseRequest, **kwargs: object) -> "Deferred[None]": d: "Deferred[None]" = Deferred() - d.addCallback(_throw) # type: ignore[unused-awaitable] + d.addCallback(_throw) # type: ignore[unused-awaitable] self.reactor.callLater(0.5, d.callback, True) return make_deferred_yieldable(d) diff --git a/tests/unittest.py b/tests/unittest.py
index 381b5adf05..532b92d43a 100644 --- a/tests/unittest.py +++ b/tests/unittest.py
@@ -555,7 +555,7 @@ class HomeserverTestCase(TestCase): deferred: Deferred[TV] = ensureDeferred(d) # type: ignore[arg-type] results: list = [] - deferred.addBoth(results.append) # type: ignore[unused-awaitable] + deferred.addBoth(results.append) # type: ignore[unused-awaitable] self.pump(by=by)