diff options
author | Brendan Abolivier <babolivier@matrix.org> | 2020-08-24 18:21:04 +0100 |
---|---|---|
committer | GitHub <noreply@github.com> | 2020-08-24 18:21:04 +0100 |
commit | 420484a334a79b31e689bdcca2e57d9a23f7e3d4 (patch) | |
tree | 36ab7908cfd72ba0266834947dcee159c3af0c5e /docs | |
parent | Add healthcheck for default localhost 8008 port on /health endpoint. (#8147) (diff) | |
download | synapse-420484a334a79b31e689bdcca2e57d9a23f7e3d4.tar.xz |
Allow capping a room's retention policy (#8104)
Diffstat (limited to 'docs')
-rw-r--r-- | docs/sample_config.yaml | 22 |
1 files changed, 14 insertions, 8 deletions
diff --git a/docs/sample_config.yaml b/docs/sample_config.yaml index f168853f67..3528d9e11f 100644 --- a/docs/sample_config.yaml +++ b/docs/sample_config.yaml @@ -378,11 +378,10 @@ retention: # min_lifetime: 1d # max_lifetime: 1y - # Retention policy limits. If set, a user won't be able to send a - # 'm.room.retention' event which features a 'min_lifetime' or a 'max_lifetime' - # that's not within this range. This is especially useful in closed federations, - # in which server admins can make sure every federating server applies the same - # rules. + # Retention policy limits. If set, and the state of a room contains a + # 'm.room.retention' event in its state which contains a 'min_lifetime' or a + # 'max_lifetime' that's out of these bounds, Synapse will cap the room's policy + # to these limits when running purge jobs. # #allowed_lifetime_min: 1d #allowed_lifetime_max: 1y @@ -408,12 +407,19 @@ retention: # (e.g. every 12h), but not want that purge to be performed by a job that's # iterating over every room it knows, which could be heavy on the server. # + # If any purge job is configured, it is strongly recommended to have at least + # a single job with neither 'shortest_max_lifetime' nor 'longest_max_lifetime' + # set, or one job without 'shortest_max_lifetime' and one job without + # 'longest_max_lifetime' set. Otherwise some rooms might be ignored, even if + # 'allowed_lifetime_min' and 'allowed_lifetime_max' are set, because capping a + # room's policy to these values is done after the policies are retrieved from + # Synapse's database (which is done using the range specified in a purge job's + # configuration). + # #purge_jobs: - # - shortest_max_lifetime: 1d - # longest_max_lifetime: 3d + # - longest_max_lifetime: 3d # interval: 12h # - shortest_max_lifetime: 3d - # longest_max_lifetime: 1y # interval: 1d # Inhibits the /requestToken endpoints from returning an error that might leak |