summary refs log tree commit diff
diff options
context:
space:
mode:
authorMark Haines <mark.haines@matrix.org>2014-12-16 17:29:49 +0000
committerMark Haines <mark.haines@matrix.org>2014-12-16 17:29:49 +0000
commit56db465047c2906a78eccdd63b2fa18d72903a66 (patch)
treebdd082c04fd0c02b4d732bf8ea6e4dd559d84b51
parentSYN-142: Use a default log file 'homeserver.log' so people get logging by def... (diff)
parentChange upgrade script to not check hashes or signatures (diff)
downloadsynapse-56db465047c2906a78eccdd63b2fa18d72903a66.tar.xz
Merge branch 'release-v0.6.0' into develop
-rw-r--r--CHANGES.rst9
-rw-r--r--UPGRADE.rst16
-rw-r--r--VERSION2
-rw-r--r--scripts/upgrade_db_to_v0.6.0.py (renamed from scripts/upgrade_db_to_v0.5.5.py)121
-rw-r--r--synapse/__init__.py2
-rwxr-xr-xsynapse/app/homeserver.py13
-rw-r--r--synapse/handlers/message.py5
-rw-r--r--synapse/state.py7
-rw-r--r--synapse/storage/__init__.py14
-rw-r--r--synapse/storage/schema/delta/v9.sql58
10 files changed, 194 insertions, 53 deletions
diff --git a/CHANGES.rst b/CHANGES.rst
index 0d36e8eeff..23bdac6a8a 100644
--- a/CHANGES.rst
+++ b/CHANGES.rst
@@ -1,3 +1,12 @@
+Changes in synapse 0.6.0 (2014-12-16)
+=====================================
+
+ * Add new API for media upload and download that supports thumbnailing.
+ * Implement typing notifications.
+ * Fix bugs where we sent events with invalid signatures due to bugs where
+   we incorrectly persisted events.
+ * Improve performance of database queries involving retrieving events.
+
 Changes in synapse 0.5.4a (2014-12-13)
 ======================================
 
diff --git a/UPGRADE.rst b/UPGRADE.rst
index 5ebdd455c1..a602a9f3eb 100644
--- a/UPGRADE.rst
+++ b/UPGRADE.rst
@@ -1,3 +1,19 @@
+Upgrading to v0.6.0
+===================
+
+This update includes a change to the database schema. To upgrade you first need
+to upgrade the database by running::
+
+    python scripts/upgrade_db_to_v0.6.0.py <db> <server_name> <signing_key>
+
+Where `<db>` is the location of the database, `<server_name>` is the
+server name as specified in the synapse configuration, and `<signing_key>` is
+the location of the signing key as specified in the synapse configuration.
+
+This may take some time to complete. Failures of signatures and content hashes
+can safely be ignored.
+
+
 Upgrading to v0.5.1
 ===================
 
diff --git a/VERSION b/VERSION
index 64c6e2f460..a918a2aa18 100644
--- a/VERSION
+++ b/VERSION
@@ -1 +1 @@
-0.5.4a
+0.6.0
diff --git a/scripts/upgrade_db_to_v0.5.5.py b/scripts/upgrade_db_to_v0.6.0.py
index aa0d7667ac..298e9c0789 100644
--- a/scripts/upgrade_db_to_v0.5.5.py
+++ b/scripts/upgrade_db_to_v0.6.0.py
@@ -1,3 +1,5 @@
+
+from synapse.storage import SCHEMA_VERSION, read_schema
 from synapse.storage._base import SQLBaseStore
 from synapse.storage.signatures import SignatureStore
 from synapse.storage.event_federation import EventFederationStore
@@ -19,8 +21,9 @@ from syutil.crypto.signing_key import decode_verify_key_bytes
 from syutil.jsonutil import encode_canonical_json
 
 import argparse
-import dns.resolver
+# import dns.resolver
 import hashlib
+import httplib
 import json
 import sqlite3
 import syutil
@@ -38,6 +41,8 @@ CREATE TABLE IF NOT EXISTS event_json(
 
 CREATE INDEX IF NOT EXISTS event_json_id ON event_json(event_id);
 CREATE INDEX IF NOT EXISTS event_json_room_id ON event_json(room_id);
+
+PRAGMA user_version = 10;
 """
 
 
@@ -142,50 +147,57 @@ class Store(object):
 store = Store()
 
 
-def get_key(server_name):
-    print "Getting keys for: %s" % (server_name,)
-    targets = []
-    if ":" in server_name:
-        target, port = server_name.split(":")
-        targets.append((target, int(port)))
-        return
-    try:
-        answers = dns.resolver.query("_matrix._tcp." + server_name, "SRV")
-        for srv in answers:
-            targets.append((srv.target, srv.port))
-    except dns.resolver.NXDOMAIN:
-        targets.append((server_name, 8448))
-    except:
-        print "Failed to lookup keys for %s" % (server_name,)
-        return {}
-
-    for target, port in targets:
-        url = "https://%s:%i/_matrix/key/v1" % (target, port)
-        try:
-            keys = json.load(urllib2.urlopen(url, timeout=2))
-            verify_keys = {}
-            for key_id, key_base64 in keys["verify_keys"].items():
-                verify_key = decode_verify_key_bytes(
-                    key_id, decode_base64(key_base64)
-                )
-                verify_signed_json(keys, server_name, verify_key)
-                verify_keys[key_id] = verify_key
-            print "Got keys for: %s" % (server_name,)
-            return verify_keys
-        except urllib2.URLError:
-            pass
-
-    print "Failed to get keys for %s" % (server_name,)
-    return {}
+# def get_key(server_name):
+#     print "Getting keys for: %s" % (server_name,)
+#     targets = []
+#     if ":" in server_name:
+#         target, port = server_name.split(":")
+#         targets.append((target, int(port)))
+#     try:
+#         answers = dns.resolver.query("_matrix._tcp." + server_name, "SRV")
+#         for srv in answers:
+#             targets.append((srv.target, srv.port))
+#     except dns.resolver.NXDOMAIN:
+#         targets.append((server_name, 8448))
+#     except:
+#         print "Failed to lookup keys for %s" % (server_name,)
+#         return {}
+#
+#     for target, port in targets:
+#         url = "https://%s:%i/_matrix/key/v1" % (target, port)
+#         try:
+#             keys = json.load(urllib2.urlopen(url, timeout=2))
+#             verify_keys = {}
+#             for key_id, key_base64 in keys["verify_keys"].items():
+#                 verify_key = decode_verify_key_bytes(
+#                     key_id, decode_base64(key_base64)
+#                 )
+#                 verify_signed_json(keys, server_name, verify_key)
+#                 verify_keys[key_id] = verify_key
+#             print "Got keys for: %s" % (server_name,)
+#             return verify_keys
+#         except urllib2.URLError:
+#             pass
+#         except urllib2.HTTPError:
+#             pass
+#         except httplib.HTTPException:
+#             pass
+#
+#     print "Failed to get keys for %s" % (server_name,)
+#     return {}
 
 
 def reinsert_events(cursor, server_name, signing_key):
+    print "Running delta: v10"
+
     cursor.executescript(delta_sql)
 
     cursor.execute(
         "SELECT * FROM events ORDER BY rowid ASC"
     )
 
+    print "Getting events..."
+
     rows = store.cursor_to_dict(cursor)
 
     events = store._generate_event_json(cursor, rows)
@@ -207,13 +219,20 @@ def reinsert_events(cursor, server_name, signing_key):
         }
     }
 
+    i = 0
+    N = len(events)
+
     for event in events:
-        for alg_name in event.hashes:
-            if check_event_content_hash(event, algorithms[alg_name]):
-                pass
-            else:
-                pass
-                print "FAIL content hash %s %s" % (alg_name, event.event_id, )
+        if i % 100 == 0:
+            print "Processed: %d/%d events" % (i,N,)
+        i += 1
+
+        # for alg_name in event.hashes:
+        #     if check_event_content_hash(event, algorithms[alg_name]):
+        #         pass
+        #     else:
+        #         pass
+        #         print "FAIL content hash %s %s" % (alg_name, event.event_id, )
 
         have_own_correctly_signed = False
         for host, sigs in event.signatures.items():
@@ -221,7 +240,7 @@ def reinsert_events(cursor, server_name, signing_key):
 
             for key_id in sigs:
                 if host not in server_keys:
-                    server_keys[host] = get_key(host)
+                    server_keys[host] = {}  # get_key(host)
                 if key_id in server_keys[host]:
                     try:
                         verify_signed_json(
@@ -275,9 +294,25 @@ def reinsert_events(cursor, server_name, signing_key):
 def main(database, server_name, signing_key):
     conn = sqlite3.connect(database)
     cursor = conn.cursor()
+
+    # Do other deltas:
+    cursor.execute("PRAGMA user_version")
+    row = cursor.fetchone()
+
+    if row and row[0]:
+        user_version = row[0]
+        # Run every version since after the current version.
+        for v in range(user_version + 1, 10):
+            print "Running delta: %d" % (v,)
+            sql_script = read_schema("delta/v%d" % (v,))
+            cursor.executescript(sql_script)
+
     reinsert_events(cursor, server_name, signing_key)
+
     conn.commit()
 
+    print "Success!"
+
 
 if __name__ == "__main__":
     parser = argparse.ArgumentParser()
diff --git a/synapse/__init__.py b/synapse/__init__.py
index 7e49e1fd08..1cef40925f 100644
--- a/synapse/__init__.py
+++ b/synapse/__init__.py
@@ -16,4 +16,4 @@
 """ This is a reference implementation of a synapse home server.
 """
 
-__version__ = "0.5.4a"
+__version__ = "0.6.0"
diff --git a/synapse/app/homeserver.py b/synapse/app/homeserver.py
index a6e29c0860..140c99f18a 100755
--- a/synapse/app/homeserver.py
+++ b/synapse/app/homeserver.py
@@ -14,7 +14,7 @@
 # See the License for the specific language governing permissions and
 # limitations under the License.
 
-from synapse.storage import prepare_database
+from synapse.storage import prepare_database, UpgradeDatabaseException
 
 from synapse.server import HomeServer
 
@@ -228,8 +228,15 @@ def setup():
 
     logger.info("Preparing database: %s...", db_name)
 
-    with sqlite3.connect(db_name) as db_conn:
-        prepare_database(db_conn)
+    try:
+        with sqlite3.connect(db_name) as db_conn:
+            prepare_database(db_conn)
+    except UpgradeDatabaseException:
+        sys.stderr.write(
+            "\nFailed to upgrade database.\n"
+            "Have you followed any instructions in UPGRADES.rst?\n"
+        )
+        sys.exit(1)
 
     logger.info("Database prepared in %s.", db_name)
 
diff --git a/synapse/handlers/message.py b/synapse/handlers/message.py
index b529d890bb..49c0e98113 100644
--- a/synapse/handlers/message.py
+++ b/synapse/handlers/message.py
@@ -123,6 +123,11 @@ class MessageHandler(BaseHandler):
 
         self.validator.validate_new(builder)
 
+        self.ratelimit(builder.user_id)
+        # TODO(paul): Why does 'event' not have a 'user' object?
+        user = self.hs.parse_userid(builder.user_id)
+        assert self.hs.is_mine(user), "User must be our own: %s" % (user,)
+
         if builder.type == EventTypes.Member:
             membership = builder.content.get("membership", None)
             if membership == Membership.JOIN:
diff --git a/synapse/state.py b/synapse/state.py
index 551058b510..15397e2f38 100644
--- a/synapse/state.py
+++ b/synapse/state.py
@@ -203,9 +203,10 @@ class StateHandler(object):
         }
 
         if event_type:
-            prev_states = conflicted_state.get(
-                (event_type, state_key), {}
-            ).keys()
+            prev_states_events = conflicted_state.get(
+                (event_type, state_key), []
+            )
+            prev_states = [s.event_id for s in prev_states_events]
         else:
             prev_states = []
 
diff --git a/synapse/storage/__init__.py b/synapse/storage/__init__.py
index c051f33a5c..60c2d67425 100644
--- a/synapse/storage/__init__.py
+++ b/synapse/storage/__init__.py
@@ -66,7 +66,7 @@ SCHEMAS = [
 
 # Remember to update this number every time an incompatible change is made to
 # database schema files, so the users will be informed on server restarts.
-SCHEMA_VERSION = 9
+SCHEMA_VERSION = 10
 
 
 class _RollbackButIsFineException(Exception):
@@ -446,6 +446,14 @@ def read_schema(schema):
         return schema_file.read()
 
 
+class PrepareDatabaseException(Exception):
+    pass
+
+
+class UpgradeDatabaseException(PrepareDatabaseException):
+    pass
+
+
 def prepare_database(db_conn):
     """ Set up all the dbs. Since all the *.sql have IF NOT EXISTS, so we
     don't have to worry about overwriting existing content.
@@ -470,6 +478,10 @@ def prepare_database(db_conn):
 
             # Run every version since after the current version.
             for v in range(user_version + 1, SCHEMA_VERSION + 1):
+                if v == 10:
+                    raise UpgradeDatabaseException(
+                        "No delta for version 10"
+                    )
                 sql_script = read_schema("delta/v%d" % (v))
                 c.executescript(sql_script)
 
diff --git a/synapse/storage/schema/delta/v9.sql b/synapse/storage/schema/delta/v9.sql
index ad680c64da..0af29733a0 100644
--- a/synapse/storage/schema/delta/v9.sql
+++ b/synapse/storage/schema/delta/v9.sql
@@ -20,4 +20,60 @@ CREATE TABLE IF NOT EXISTS destinations(
     retry_interval INTEGER
 );
 
-PRAGMA user_version = 9;
\ No newline at end of file
+
+CREATE TABLE IF NOT EXISTS local_media_repository (
+    media_id TEXT, -- The id used to refer to the media.
+    media_type TEXT, -- The MIME-type of the media.
+    media_length INTEGER, -- Length of the media in bytes.
+    created_ts INTEGER, -- When the content was uploaded in ms.
+    upload_name TEXT, -- The name the media was uploaded with.
+    user_id TEXT, -- The user who uploaded the file.
+    CONSTRAINT uniqueness UNIQUE (media_id)
+);
+
+CREATE TABLE IF NOT EXISTS local_media_repository_thumbnails (
+    media_id TEXT, -- The id used to refer to the media.
+    thumbnail_width INTEGER, -- The width of the thumbnail in pixels.
+    thumbnail_height INTEGER, -- The height of the thumbnail in pixels.
+    thumbnail_type TEXT, -- The MIME-type of the thumbnail.
+    thumbnail_method TEXT, -- The method used to make the thumbnail.
+    thumbnail_length INTEGER, -- The length of the thumbnail in bytes.
+    CONSTRAINT uniqueness UNIQUE (
+        media_id, thumbnail_width, thumbnail_height, thumbnail_type
+    )
+);
+
+CREATE INDEX IF NOT EXISTS local_media_repository_thumbnails_media_id
+    ON local_media_repository_thumbnails (media_id);
+
+CREATE TABLE IF NOT EXISTS remote_media_cache (
+    media_origin TEXT, -- The remote HS the media came from.
+    media_id TEXT, -- The id used to refer to the media on that server.
+    media_type TEXT, -- The MIME-type of the media.
+    created_ts INTEGER, -- When the content was uploaded in ms.
+    upload_name TEXT, -- The name the media was uploaded with.
+    media_length INTEGER, -- Length of the media in bytes.
+    filesystem_id TEXT, -- The name used to store the media on disk.
+    CONSTRAINT uniqueness UNIQUE (media_origin, media_id)
+);
+
+CREATE TABLE IF NOT EXISTS remote_media_cache_thumbnails (
+    media_origin TEXT, -- The remote HS the media came from.
+    media_id TEXT, -- The id used to refer to the media.
+    thumbnail_width INTEGER, -- The width of the thumbnail in pixels.
+    thumbnail_height INTEGER, -- The height of the thumbnail in pixels.
+    thumbnail_method TEXT, -- The method used to make the thumbnail
+    thumbnail_type TEXT, -- The MIME-type of the thumbnail.
+    thumbnail_length INTEGER, -- The length of the thumbnail in bytes.
+    filesystem_id TEXT, -- The name used to store the media on disk.
+    CONSTRAINT uniqueness UNIQUE (
+        media_origin, media_id, thumbnail_width, thumbnail_height,
+        thumbnail_type, thumbnail_type
+    )
+);
+
+CREATE INDEX IF NOT EXISTS remote_media_cache_thumbnails_media_id
+    ON local_media_repository_thumbnails (media_id);
+
+
+PRAGMA user_version = 9;