summary refs log tree commit diff
path: root/synapse/storage
diff options
context:
space:
mode:
authorErik Johnston <erik@matrix.org>2016-08-15 11:16:45 +0100
committerErik Johnston <erik@matrix.org>2016-08-15 11:16:45 +0100
commit64e7e1185392972fd85718bfa55248b041f56b82 (patch)
treeb41fae09aaaa891ede5c06ecc84db633403e4985 /synapse/storage
parentAdd some invalidations to a cache_stream (diff)
downloadsynapse-64e7e1185392972fd85718bfa55248b041f56b82.tar.xz
Implement cache replication stream
Diffstat (limited to 'synapse/storage')
-rw-r--r--synapse/storage/__init__.py11
-rw-r--r--synapse/storage/_base.py47
2 files changed, 43 insertions, 15 deletions
diff --git a/synapse/storage/__init__.py b/synapse/storage/__init__.py
index a0c029a2fc..8af492b69f 100644
--- a/synapse/storage/__init__.py
+++ b/synapse/storage/__init__.py
@@ -50,6 +50,7 @@ from .openid import OpenIdStore
 from .client_ips import ClientIpStore
 
 from .util.id_generators import IdGenerator, StreamIdGenerator, ChainedIdGenerator
+from .engines import PostgresEngine
 
 from synapse.api.constants import PresenceState
 from synapse.util.caches.stream_change_cache import StreamChangeCache
@@ -122,9 +123,13 @@ class DataStore(RoomMemberStore, RoomStore,
             db_conn, "pushers", "id",
             extra_tables=[("deleted_pushers", "stream_id")],
         )
-        self._cache_id_gen = StreamIdGenerator(
-            db_conn, "cache_stream", "stream_id",
-        )
+
+        if isinstance(self.database_engine, PostgresEngine):
+            self._cache_id_gen = StreamIdGenerator(
+                db_conn, "cache_stream", "stream_id",
+            )
+        else:
+            self._cache_id_gen = None
 
         events_max = self._stream_id_gen.get_current_token()
         event_cache_prefill, min_event_val = self._get_cache_dict(
diff --git a/synapse/storage/_base.py b/synapse/storage/_base.py
index 02d9098ddd..e3edc2cde6 100644
--- a/synapse/storage/_base.py
+++ b/synapse/storage/_base.py
@@ -19,6 +19,7 @@ from synapse.util.logcontext import LoggingContext, PreserveLoggingContext
 from synapse.util.caches.dictionary_cache import DictionaryCache
 from synapse.util.caches.descriptors import Cache
 from synapse.util.caches import intern_dict
+from synapse.storage.engines import PostgresEngine
 import synapse.metrics
 
 
@@ -864,21 +865,43 @@ class SQLBaseStore(object):
     def _invalidate_cache_and_stream(self, txn, cache_func, keys):
         txn.call_after(cache_func.invalidate, keys)
 
-        ctx = self._cache_id_gen.get_next()
-        stream_id = ctx.__enter__()
-        txn.call_after(ctx.__exit__, None, None, None)
+        if isinstance(self.database_engine, PostgresEngine):
+            ctx = self._cache_id_gen.get_next()
+            stream_id = ctx.__enter__()
+            txn.call_after(ctx.__exit__, None, None, None)
+
+            self._simple_insert_txn(
+                txn,
+                table="cache_stream",
+                values={
+                    "stream_id": stream_id,
+                    "cache_func": cache_func.__name__,
+                    "keys": list(keys),
+                    "invalidation_ts": self.clock.time_msec(),
+                }
+            )
 
-        self._simple_insert_txn(
-            txn,
-            table="cache_stream",
-            values={
-                "stream_id": stream_id,
-                "cache_func": cache_func.__name__,
-                "keys": list(keys),
-                "invalidation_ts": self.clock.time_msec(),
-            }
+    def get_all_updated_caches(self, last_id, current_id, limit):
+        def get_all_updated_caches_txn(txn):
+            # We purposefully don't bound by the current token, as we want to
+            # send across cache invalidations as quickly as possible. Cache
+            # invalidations are idempotent, so duplicates are fine.
+            sql = (
+                "SELECT stream_id, cache_func, keys, invalidation_ts FROM cache_stream"
+                " WHERE stream_id > ? ORDER BY stream_id ASC LIMIT ?"
+            )
+            txn.execute(sql, (last_id, limit,))
+            return txn.fetchall()
+        return self.runInteraction(
+            "get_all_updated_caches", get_all_updated_caches_txn
         )
 
+    def get_cache_stream_token(self):
+        if self._cache_id_gen:
+            return self._cache_id_gen.get_current_token()
+        else:
+            return 0
+
 
 class _RollbackButIsFineException(Exception):
     """ This exception is used to rollback a transaction without implying