diff --git a/synapse/http/client.py b/synapse/http/client.py
index 3b8ffcd3ef..6c89b20984 100644
--- a/synapse/http/client.py
+++ b/synapse/http/client.py
@@ -271,12 +271,19 @@ class SimpleHttpClient(object):
if 'Content-Length' in headers and headers['Content-Length'] > max_size:
logger.warn("Requested URL is too large > %r bytes" % (self.max_size,))
- # XXX: do we want to explicitly drop the connection here somehow? if so, how?
- raise # what should we be raising here?
+ raise SynapseError(
+ 502,
+ "Requested file is too large > %r bytes" % (self.max_size,),
+ Codes.TOO_LARGE,
+ )
if response.code > 299:
logger.warn("Got %d when downloading %s" % (response.code, url))
- raise
+ raise SynapseError(
+ 502,
+ "Got error %d" % (response.code,),
+ Codes.UNKNOWN,
+ )
# TODO: if our Content-Type is HTML or something, just read the first
# N bytes into RAM rather than saving it all to disk only to read it
@@ -287,9 +294,13 @@ class SimpleHttpClient(object):
_readBodyToFile,
response, output_stream, max_size
)
- except:
+ except Exception as e:
logger.exception("Failed to download body")
- raise
+ raise SynapseError(
+ 502,
+ ("Failed to download remote body: %s" % e),
+ Codes.UNKNOWN,
+ )
defer.returnValue((length, headers, response.request.absoluteURI, response.code))
diff --git a/synapse/rest/media/v1/preview_url_resource.py b/synapse/rest/media/v1/preview_url_resource.py
index 4df9099499..c72c73ca82 100644
--- a/synapse/rest/media/v1/preview_url_resource.py
+++ b/synapse/rest/media/v1/preview_url_resource.py
@@ -19,6 +19,9 @@ from twisted.web.server import NOT_DONE_YET
from twisted.internet import defer
from urlparse import urlparse, urlsplit, urlunparse
+from synapse.api.errors import (
+ SynapseError, Codes,
+)
from synapse.util.stringutils import random_string
from synapse.util.caches.expiringcache import ExpiringCache
from synapse.http.client import SpiderHttpClient
@@ -47,9 +50,11 @@ class PreviewUrlResource(BaseMediaResource):
isLeaf = True
def __init__(self, hs, filepaths):
- if not html:
- logger.warn("Disabling PreviewUrlResource as lxml not available")
- raise
+ try:
+ if html:
+ pass
+ except:
+ raise RunTimeError("Disabling PreviewUrlResource as lxml not available")
if not hasattr(hs.config, "url_preview_ip_range_blacklist"):
logger.warn(
@@ -57,7 +62,10 @@ class PreviewUrlResource(BaseMediaResource):
"blacklist in url_preview_ip_range_blacklist for url previewing "
"to work"
)
- raise
+ raise RunTimeError(
+ "Disabling PreviewUrlResource as "
+ "url_preview_ip_range_blacklist not specified"
+ )
BaseMediaResource.__init__(self, hs, filepaths)
self.client = SpiderHttpClient(hs)
@@ -121,7 +129,10 @@ class PreviewUrlResource(BaseMediaResource):
logger.warn(
"URL %s blocked by url_blacklist entry %s", url, entry
)
- raise
+ raise SynapseError(
+ 403, "URL blocked by url pattern blacklist entry",
+ Codes.UNKNOWN
+ )
# first check the memory cache - good to handle all the clients on this
# HS thundering away to preview the same URL at the same time.
@@ -229,8 +240,9 @@ class PreviewUrlResource(BaseMediaResource):
)
respond_with_json_bytes(request, 200, json.dumps(og), send_cors=True)
- except:
- raise
+ except Exception as e:
+ raise e
+
@defer.inlineCallbacks
def _calc_og(self, tree, media_info, requester):
@@ -418,9 +430,12 @@ class PreviewUrlResource(BaseMediaResource):
user_id=user,
)
- except:
+ except Exception as e:
os.remove(fname)
- raise
+ raise SynapseError(
+ 500, ("Failed to download content: %s" % e),
+ Codes.UNKNOWN
+ )
defer.returnValue({
"media_type": media_type,
|