diff options
author | Erik Johnston <erikj@element.io> | 2024-04-08 14:25:28 +0100 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-04-08 14:25:28 +0100 |
commit | 1f8f991d51a3311d67ea0b717bf168553d51b441 (patch) | |
tree | f1410e259a4c91981430781d08ad8c94bf8b6fee /synapse/handlers | |
parent | Pull out fewer receipts from DB when doing push (#17049) (diff) | |
download | synapse-1f8f991d51a3311d67ea0b717bf168553d51b441.tar.xz |
Add back fast path for non-gappy syncs (#17064)
PR #16942 removed an invalid optimisation that avoided pulling out state for non-gappy syncs. This causes a large increase in DB usage. c.f. #16941 for why that optimisation was wrong. However, we can still optimise in the simple case where the events in the timeline are a linear chain without any branching/merging of the DAG. cc. @richvdh
Diffstat (limited to 'synapse/handlers')
-rw-r--r-- | synapse/handlers/sync.py | 36 |
1 files changed, 36 insertions, 0 deletions
diff --git a/synapse/handlers/sync.py b/synapse/handlers/sync.py index 554c820f79..7c29c15540 100644 --- a/synapse/handlers/sync.py +++ b/synapse/handlers/sync.py @@ -1259,6 +1259,42 @@ class SyncHandler: await_full_state = True lazy_load_members = False + # For a non-gappy sync if the events in the timeline are simply a linear + # chain (i.e. no merging/branching of the graph), then we know the state + # delta between the end of the previous sync and start of the new one is + # empty. + # + # c.f. #16941 for an example of why we can't do this for all non-gappy + # syncs. + is_linear_timeline = False + if batch.events: + prev_event_id = batch.events[0].event_id + for e in batch.events[1:]: + if e.prev_event_ids() != [prev_event_id]: + break + else: + is_linear_timeline = True + + if is_linear_timeline and not batch.limited: + state_ids: StateMap[str] = {} + if lazy_load_members: + if members_to_fetch and batch.events: + # We're lazy-loading, so the client might need some more + # member events to understand the events in this timeline. + # So we fish out all the member events corresponding to the + # timeline here. The caller will then dedupe any redundant + # ones. + + state_ids = await self._state_storage_controller.get_state_ids_for_event( + batch.events[0].event_id, + # we only want members! + state_filter=StateFilter.from_types( + (EventTypes.Member, member) for member in members_to_fetch + ), + await_full_state=False, + ) + return state_ids + if batch: state_at_timeline_start = ( await self._state_storage_controller.get_state_ids_for_event( |