summary refs log tree commit diff
path: root/synapse/storage/stats.py
diff options
context:
space:
mode:
Diffstat (limited to 'synapse/storage/stats.py')
-rw-r--r--synapse/storage/stats.py48
1 files changed, 25 insertions, 23 deletions
diff --git a/synapse/storage/stats.py b/synapse/storage/stats.py

index 0ad40cf226..b6959f7967 100644 --- a/synapse/storage/stats.py +++ b/synapse/storage/stats.py
@@ -140,7 +140,7 @@ class StatsStore(StateDeltasStore): if not self.stats_enabled: yield self._end_background_update("populate_stats_prepare") - defer.returnValue(1) + return 1 def _wedge_incremental_processor(txn): """ @@ -218,7 +218,7 @@ class StatsStore(StateDeltasStore): self.get_earliest_token_for_stats.invalidate_all() yield self._end_background_update("populate_stats_prepare") - defer.returnValue(1) + return 1 @defer.inlineCallbacks def _populate_stats_process_users(self, progress, batch_size): @@ -227,7 +227,7 @@ class StatsStore(StateDeltasStore): """ if not self.stats_enabled: yield self._end_background_update("populate_stats_process_users") - defer.returnValue(1) + return 1 def _get_next_batch(txn): # Only fetch 250 users, so we don't fetch too many at once, even @@ -260,7 +260,7 @@ class StatsStore(StateDeltasStore): # No more users -- complete the transaction. if not users_to_work_on: yield self._end_background_update("populate_stats_process_users") - defer.returnValue(1) + return 1 logger.info( "Processing the next %d users of %d remaining", @@ -280,7 +280,7 @@ class StatsStore(StateDeltasStore): promised_positions, ) yield self._end_background_update("populate_stats_process_users") - defer.returnValue(1) + return 1 for (user_id,) in users_to_work_on: now = self.hs.get_reactor().seconds() @@ -348,7 +348,7 @@ class StatsStore(StateDeltasStore): if processed_membership_count > batch_size: # Don't process any more users, we've hit our batch size. - defer.returnValue(processed_membership_count) + return processed_membership_count yield self.runInteraction( "populate_stats", @@ -357,7 +357,7 @@ class StatsStore(StateDeltasStore): progress, ) - defer.returnValue(processed_membership_count) + return processed_membership_count @defer.inlineCallbacks def _populate_stats_process_rooms(self, progress, batch_size): @@ -366,7 +366,7 @@ class StatsStore(StateDeltasStore): """ if not self.stats_enabled: yield self._end_background_update("populate_stats_process_rooms") - defer.returnValue(1) + return 1 def _get_next_batch(txn): # Only fetch 250 rooms, so we don't fetch too many at once, even @@ -399,7 +399,7 @@ class StatsStore(StateDeltasStore): # No more rooms -- complete the transaction. if not rooms_to_work_on: yield self._end_background_update("populate_stats_process_rooms") - defer.returnValue(1) + return 1 logger.info( "Processing the next %d rooms of %d remaining", @@ -420,7 +420,7 @@ class StatsStore(StateDeltasStore): promised_positions, ) yield self._end_background_update("populate_stats_process_rooms") - defer.returnValue(1) + return 1 for (room_id,) in rooms_to_work_on: current_state_ids = yield self.get_current_state_ids(room_id) @@ -435,16 +435,18 @@ class StatsStore(StateDeltasStore): avatar_id = current_state_ids.get((EventTypes.RoomAvatar, "")) canonical_alias_id = current_state_ids.get((EventTypes.CanonicalAlias, "")) + event_ids = [ + join_rules_id, + history_visibility_id, + encryption_id, + name_id, + topic_id, + avatar_id, + canonical_alias_id, + ] + state_events = yield self.get_events( - [ - join_rules_id, - history_visibility_id, - encryption_id, - name_id, - topic_id, - avatar_id, - canonical_alias_id, - ] + [ev for ev in event_ids if ev is not None] ) def _get_or_none(event_id, arg): @@ -535,7 +537,7 @@ class StatsStore(StateDeltasStore): if processed_event_count > batch_size: # Don't process any more rooms, we've hit our batch size. - defer.returnValue(processed_event_count) + return processed_event_count yield self.runInteraction( "populate_stats", @@ -544,7 +546,7 @@ class StatsStore(StateDeltasStore): progress, ) - defer.returnValue(processed_event_count) + return processed_event_count def update_total_event_count_between_txn(self, txn, low_pos, high_pos): """ @@ -934,7 +936,7 @@ class StatsStore(StateDeltasStore): "stats_collect_old", self._collect_old_txn, stats_type ) if not maybe_more: - defer.returnValue(None) + return None @defer.inlineCallbacks def update_stats_delta( @@ -966,7 +968,7 @@ class StatsStore(StateDeltasStore): fields, complete_with_stream_id=complete_with_stream_id, ) - defer.returnValue(res) + return res except OldCollectionRequired: # retry after collecting old rows yield self.collect_old(stats_type)