mirror of
https://github.com/element-hq/synapse.git
synced 2025-12-07 01:20:16 +00:00
Compare commits
2 Commits
release-v1
...
erikj/file
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
9854f4c7ff | ||
|
|
518b3a3f89 |
@@ -164,6 +164,101 @@ class MessageHandler(BaseHandler):
|
|||||||
|
|
||||||
defer.returnValue(chunk)
|
defer.returnValue(chunk)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def get_files(self, requester, room_id, pagin_config):
|
||||||
|
"""Get files in a room.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
requester (Requester): The user requesting files.
|
||||||
|
room_id (str): The room they want files from.
|
||||||
|
pagin_config (synapse.api.streams.PaginationConfig): The pagination
|
||||||
|
config rules to apply, if any.
|
||||||
|
as_client_event (bool): True to get events in client-server format.
|
||||||
|
Returns:
|
||||||
|
dict: Pagination API results
|
||||||
|
"""
|
||||||
|
user_id = requester.user.to_string()
|
||||||
|
|
||||||
|
if pagin_config.from_token:
|
||||||
|
room_token = pagin_config.from_token.room_key
|
||||||
|
else:
|
||||||
|
pagin_config.from_token = (
|
||||||
|
yield self.hs.get_event_sources().get_current_token(
|
||||||
|
direction='b'
|
||||||
|
)
|
||||||
|
)
|
||||||
|
room_token = pagin_config.from_token.room_key
|
||||||
|
|
||||||
|
room_token = RoomStreamToken.parse(room_token)
|
||||||
|
|
||||||
|
pagin_config.from_token = pagin_config.from_token.copy_and_replace(
|
||||||
|
"room_key", str(room_token)
|
||||||
|
)
|
||||||
|
|
||||||
|
source_config = pagin_config.get_source_config("room")
|
||||||
|
|
||||||
|
membership, member_event_id = yield self._check_in_room_or_world_readable(
|
||||||
|
room_id, user_id
|
||||||
|
)
|
||||||
|
|
||||||
|
if source_config.direction == 'b':
|
||||||
|
if room_token.topological:
|
||||||
|
max_topo = room_token.topological
|
||||||
|
else:
|
||||||
|
max_topo = yield self.store.get_max_topological_token(
|
||||||
|
room_id, room_token.stream
|
||||||
|
)
|
||||||
|
|
||||||
|
if membership == Membership.LEAVE:
|
||||||
|
# If they have left the room then clamp the token to be before
|
||||||
|
# they left the room, to save the effort of loading from the
|
||||||
|
# database.
|
||||||
|
leave_token = yield self.store.get_topological_token_for_event(
|
||||||
|
member_event_id
|
||||||
|
)
|
||||||
|
leave_token = RoomStreamToken.parse(leave_token)
|
||||||
|
if leave_token.topological < max_topo:
|
||||||
|
source_config.from_key = str(leave_token)
|
||||||
|
|
||||||
|
events, next_key = yield self.store.paginate_room_file_events(
|
||||||
|
room_id,
|
||||||
|
from_key=source_config.from_key,
|
||||||
|
to_key=source_config.to_key,
|
||||||
|
direction=source_config.direction,
|
||||||
|
limit=source_config.limit,
|
||||||
|
)
|
||||||
|
|
||||||
|
next_token = pagin_config.from_token.copy_and_replace(
|
||||||
|
"room_key", next_key
|
||||||
|
)
|
||||||
|
|
||||||
|
if not events:
|
||||||
|
defer.returnValue({
|
||||||
|
"chunk": [],
|
||||||
|
"start": pagin_config.from_token.to_string(),
|
||||||
|
"end": next_token.to_string(),
|
||||||
|
})
|
||||||
|
|
||||||
|
events = yield filter_events_for_client(
|
||||||
|
self.store,
|
||||||
|
user_id,
|
||||||
|
events,
|
||||||
|
is_peeking=(member_event_id is None),
|
||||||
|
)
|
||||||
|
|
||||||
|
time_now = self.clock.time_msec()
|
||||||
|
|
||||||
|
chunk = {
|
||||||
|
"chunk": [
|
||||||
|
serialize_event(e, time_now)
|
||||||
|
for e in events
|
||||||
|
],
|
||||||
|
"start": pagin_config.from_token.to_string(),
|
||||||
|
"end": next_token.to_string(),
|
||||||
|
}
|
||||||
|
|
||||||
|
defer.returnValue(chunk)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def create_event(self, event_dict, token_id=None, txn_id=None, prev_event_ids=None):
|
def create_event(self, event_dict, token_id=None, txn_id=None, prev_event_ids=None):
|
||||||
"""
|
"""
|
||||||
|
|||||||
@@ -338,6 +338,25 @@ class RoomMessageListRestServlet(ClientV1RestServlet):
|
|||||||
defer.returnValue((200, msgs))
|
defer.returnValue((200, msgs))
|
||||||
|
|
||||||
|
|
||||||
|
class RoomFileListRestServlet(ClientV1RestServlet):
|
||||||
|
PATTERNS = client_path_patterns("/rooms/(?P<room_id>[^/]*)/files$")
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def on_GET(self, request, room_id):
|
||||||
|
requester = yield self.auth.get_user_by_req(request, allow_guest=True)
|
||||||
|
pagination_config = PaginationConfig.from_request(
|
||||||
|
request, default_limit=10, default_dir='b',
|
||||||
|
)
|
||||||
|
handler = self.handlers.message_handler
|
||||||
|
msgs = yield handler.get_files(
|
||||||
|
room_id=room_id,
|
||||||
|
requester=requester,
|
||||||
|
pagin_config=pagination_config,
|
||||||
|
)
|
||||||
|
|
||||||
|
defer.returnValue((200, msgs))
|
||||||
|
|
||||||
|
|
||||||
# TODO: Needs unit testing
|
# TODO: Needs unit testing
|
||||||
class RoomStateRestServlet(ClientV1RestServlet):
|
class RoomStateRestServlet(ClientV1RestServlet):
|
||||||
PATTERNS = client_path_patterns("/rooms/(?P<room_id>[^/]*)/state$")
|
PATTERNS = client_path_patterns("/rooms/(?P<room_id>[^/]*)/state$")
|
||||||
@@ -667,6 +686,7 @@ def register_servlets(hs, http_server):
|
|||||||
RoomCreateRestServlet(hs).register(http_server)
|
RoomCreateRestServlet(hs).register(http_server)
|
||||||
RoomMemberListRestServlet(hs).register(http_server)
|
RoomMemberListRestServlet(hs).register(http_server)
|
||||||
RoomMessageListRestServlet(hs).register(http_server)
|
RoomMessageListRestServlet(hs).register(http_server)
|
||||||
|
RoomFileListRestServlet(hs).register(http_server)
|
||||||
JoinRoomAliasServlet(hs).register(http_server)
|
JoinRoomAliasServlet(hs).register(http_server)
|
||||||
RoomForgetRestServlet(hs).register(http_server)
|
RoomForgetRestServlet(hs).register(http_server)
|
||||||
RoomMembershipRestServlet(hs).register(http_server)
|
RoomMembershipRestServlet(hs).register(http_server)
|
||||||
|
|||||||
@@ -34,6 +34,108 @@ OpsLevel = collections.namedtuple(
|
|||||||
|
|
||||||
|
|
||||||
class RoomStore(SQLBaseStore):
|
class RoomStore(SQLBaseStore):
|
||||||
|
EVENT_FILES_UPDATE_NAME = "event_files"
|
||||||
|
|
||||||
|
FILE_MSGTYPES = (
|
||||||
|
"m.image",
|
||||||
|
"m.video",
|
||||||
|
"m.file",
|
||||||
|
"m.audio",
|
||||||
|
)
|
||||||
|
|
||||||
|
def __init__(self, hs):
|
||||||
|
super(RoomStore, self).__init__(hs)
|
||||||
|
self.register_background_update_handler(
|
||||||
|
self.EVENT_FILES_UPDATE_NAME, self._background_reindex_files
|
||||||
|
)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def _background_reindex_files(self, progress, batch_size):
|
||||||
|
target_min_stream_id = progress["target_min_stream_id_inclusive"]
|
||||||
|
max_stream_id = progress["max_stream_id_exclusive"]
|
||||||
|
rows_inserted = progress.get("rows_inserted", 0)
|
||||||
|
|
||||||
|
def reindex_txn(txn):
|
||||||
|
sql = (
|
||||||
|
"SELECT topological_ordering, stream_ordering, event_id, room_id,"
|
||||||
|
" type, content FROM events"
|
||||||
|
" WHERE ? <= stream_ordering AND stream_ordering < ?"
|
||||||
|
" AND type = 'm.room.message'"
|
||||||
|
" AND content LIKE ?"
|
||||||
|
" ORDER BY stream_ordering DESC"
|
||||||
|
" LIMIT ?"
|
||||||
|
)
|
||||||
|
|
||||||
|
txn.execute(sql, (target_min_stream_id, max_stream_id, '%url%', batch_size))
|
||||||
|
|
||||||
|
rows = self.cursor_to_dict(txn)
|
||||||
|
if not rows:
|
||||||
|
return 0
|
||||||
|
|
||||||
|
min_stream_id = rows[-1]["stream_ordering"]
|
||||||
|
|
||||||
|
event_files_rows = []
|
||||||
|
for row in rows:
|
||||||
|
try:
|
||||||
|
so = row["stream_ordering"]
|
||||||
|
to = row["topological_ordering"]
|
||||||
|
event_id = row["event_id"]
|
||||||
|
room_id = row["room_id"]
|
||||||
|
try:
|
||||||
|
content = json.loads(row["content"])
|
||||||
|
except:
|
||||||
|
continue
|
||||||
|
|
||||||
|
msgtype = content["msgtype"]
|
||||||
|
if msgtype not in self.FILE_MSGTYPES:
|
||||||
|
continue
|
||||||
|
|
||||||
|
url = content["url"]
|
||||||
|
|
||||||
|
if not isinstance(url, basestring):
|
||||||
|
continue
|
||||||
|
if not isinstance(msgtype, basestring):
|
||||||
|
continue
|
||||||
|
except (KeyError, AttributeError):
|
||||||
|
# If the event is missing a necessary field then
|
||||||
|
# skip over it.
|
||||||
|
continue
|
||||||
|
|
||||||
|
event_files_rows.append({
|
||||||
|
"topological_ordering": to,
|
||||||
|
"stream_ordering": so,
|
||||||
|
"event_id": event_id,
|
||||||
|
"room_id": room_id,
|
||||||
|
"msgtype": msgtype,
|
||||||
|
"url": url,
|
||||||
|
})
|
||||||
|
|
||||||
|
self._simple_insert_many_txn(
|
||||||
|
txn,
|
||||||
|
table="event_files",
|
||||||
|
values=event_files_rows,
|
||||||
|
)
|
||||||
|
|
||||||
|
progress = {
|
||||||
|
"target_min_stream_id_inclusive": target_min_stream_id,
|
||||||
|
"max_stream_id_exclusive": min_stream_id,
|
||||||
|
"rows_inserted": rows_inserted + len(event_files_rows)
|
||||||
|
}
|
||||||
|
|
||||||
|
self._background_update_progress_txn(
|
||||||
|
txn, self.EVENT_FILES_UPDATE_NAME, progress
|
||||||
|
)
|
||||||
|
|
||||||
|
return len(rows)
|
||||||
|
|
||||||
|
result = yield self.runInteraction(
|
||||||
|
self.EVENT_FILES_UPDATE_NAME, reindex_txn
|
||||||
|
)
|
||||||
|
|
||||||
|
if not result:
|
||||||
|
yield self._end_background_update(self.EVENT_FILES_UPDATE_NAME)
|
||||||
|
|
||||||
|
defer.returnValue(result)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def store_room(self, room_id, room_creator_user_id, is_public):
|
def store_room(self, room_id, room_creator_user_id, is_public):
|
||||||
@@ -142,6 +244,22 @@ class RoomStore(SQLBaseStore):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def _store_room_message_txn(self, txn, event):
|
def _store_room_message_txn(self, txn, event):
|
||||||
|
msgtype = event.content.get("msgtype")
|
||||||
|
url = event.content.get("url")
|
||||||
|
if msgtype in self.FILE_MSGTYPES and url:
|
||||||
|
self._simple_insert_txn(
|
||||||
|
txn,
|
||||||
|
table="event_files",
|
||||||
|
values={
|
||||||
|
"topological_ordering": event.depth,
|
||||||
|
"stream_ordering": event.internal_metadata.stream_ordering,
|
||||||
|
"room_id": event.room_id,
|
||||||
|
"event_id": event.event_id,
|
||||||
|
"msgtype": msgtype,
|
||||||
|
"url": url,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
if hasattr(event, "content") and "body" in event.content:
|
if hasattr(event, "content") and "body" in event.content:
|
||||||
self._store_event_search_txn(
|
self._store_event_search_txn(
|
||||||
txn, event, "content.body", event.content["body"]
|
txn, event, "content.body", event.content["body"]
|
||||||
|
|||||||
72
synapse/storage/schema/delta/33/msgtype.py
Normal file
72
synapse/storage/schema/delta/33/msgtype.py
Normal file
@@ -0,0 +1,72 @@
|
|||||||
|
# Copyright 2016 OpenMarket Ltd
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
|
import logging
|
||||||
|
import ujson
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
CREATE_TABLE = """
|
||||||
|
CREATE TABLE event_files(
|
||||||
|
topological_ordering BIGINT NOT NULL,
|
||||||
|
stream_ordering BIGINT NOT NULL,
|
||||||
|
room_id TEXT NOT NULL,
|
||||||
|
event_id TEXT NOT NULL,
|
||||||
|
msgtype TEXT NOT NULL,
|
||||||
|
url TEXT NOT NULL
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE INDEX event_files_rm_id ON event_files(room_id, event_id);
|
||||||
|
CREATE INDEX event_files_order ON event_files(
|
||||||
|
room_id, topological_ordering, stream_ordering
|
||||||
|
);
|
||||||
|
CREATE INDEX event_files_order_stream ON event_files(room_id, stream_ordering);
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
def run_create(cur, database_engine, *args, **kwargs):
|
||||||
|
for statement in get_statements(CREATE_TABLE.splitlines()):
|
||||||
|
cur.execute(statement)
|
||||||
|
|
||||||
|
cur.execute("SELECT MIN(stream_ordering) FROM events")
|
||||||
|
rows = cur.fetchall()
|
||||||
|
min_stream_id = rows[0][0]
|
||||||
|
|
||||||
|
cur.execute("SELECT MAX(stream_ordering) FROM events")
|
||||||
|
rows = cur.fetchall()
|
||||||
|
max_stream_id = rows[0][0]
|
||||||
|
|
||||||
|
if min_stream_id is not None and max_stream_id is not None:
|
||||||
|
progress = {
|
||||||
|
"target_min_stream_id_inclusive": min_stream_id,
|
||||||
|
"max_stream_id_exclusive": max_stream_id + 1,
|
||||||
|
"rows_inserted": 0,
|
||||||
|
}
|
||||||
|
progress_json = ujson.dumps(progress)
|
||||||
|
|
||||||
|
sql = (
|
||||||
|
"INSERT into background_updates (update_name, progress_json)"
|
||||||
|
" VALUES (?, ?)"
|
||||||
|
)
|
||||||
|
|
||||||
|
sql = database_engine.convert_param_style(sql)
|
||||||
|
|
||||||
|
cur.execute(sql, ("event_files", progress_json))
|
||||||
|
|
||||||
|
|
||||||
|
def run_upgrade(cur, database_engine, *args, **kwargs):
|
||||||
|
pass
|
||||||
@@ -394,6 +394,82 @@ class StreamStore(SQLBaseStore):
|
|||||||
|
|
||||||
defer.returnValue((events, token))
|
defer.returnValue((events, token))
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def paginate_room_file_events(self, room_id, from_key, to_key=None,
|
||||||
|
direction='b', limit=-1):
|
||||||
|
# Tokens really represent positions between elements, but we use
|
||||||
|
# the convention of pointing to the event before the gap. Hence
|
||||||
|
# we have a bit of asymmetry when it comes to equalities.
|
||||||
|
args = [room_id]
|
||||||
|
if direction == 'b':
|
||||||
|
order = "DESC"
|
||||||
|
bounds = upper_bound(
|
||||||
|
RoomStreamToken.parse(from_key), self.database_engine
|
||||||
|
)
|
||||||
|
if to_key:
|
||||||
|
bounds = "%s AND %s" % (bounds, lower_bound(
|
||||||
|
RoomStreamToken.parse(to_key), self.database_engine
|
||||||
|
))
|
||||||
|
else:
|
||||||
|
order = "ASC"
|
||||||
|
bounds = lower_bound(
|
||||||
|
RoomStreamToken.parse(from_key), self.database_engine
|
||||||
|
)
|
||||||
|
if to_key:
|
||||||
|
bounds = "%s AND %s" % (bounds, upper_bound(
|
||||||
|
RoomStreamToken.parse(to_key), self.database_engine
|
||||||
|
))
|
||||||
|
|
||||||
|
if int(limit) > 0:
|
||||||
|
args.append(int(limit))
|
||||||
|
limit_str = " LIMIT ?"
|
||||||
|
else:
|
||||||
|
limit_str = ""
|
||||||
|
|
||||||
|
sql = (
|
||||||
|
"SELECT * FROM event_files"
|
||||||
|
" WHERE room_id = ? AND %(bounds)s"
|
||||||
|
" ORDER BY topological_ordering %(order)s,"
|
||||||
|
" stream_ordering %(order)s %(limit)s"
|
||||||
|
) % {
|
||||||
|
"bounds": bounds,
|
||||||
|
"order": order,
|
||||||
|
"limit": limit_str
|
||||||
|
}
|
||||||
|
|
||||||
|
def f(txn):
|
||||||
|
txn.execute(sql, args)
|
||||||
|
|
||||||
|
rows = self.cursor_to_dict(txn)
|
||||||
|
|
||||||
|
if rows:
|
||||||
|
topo = rows[-1]["topological_ordering"]
|
||||||
|
toke = rows[-1]["stream_ordering"]
|
||||||
|
if direction == 'b':
|
||||||
|
# Tokens are positions between events.
|
||||||
|
# This token points *after* the last event in the chunk.
|
||||||
|
# We need it to point to the event before it in the chunk
|
||||||
|
# when we are going backwards so we subtract one from the
|
||||||
|
# stream part.
|
||||||
|
toke -= 1
|
||||||
|
next_token = str(RoomStreamToken(topo, toke))
|
||||||
|
else:
|
||||||
|
# TODO (erikj): We should work out what to do here instead.
|
||||||
|
next_token = to_key if to_key else from_key
|
||||||
|
|
||||||
|
return rows, next_token,
|
||||||
|
|
||||||
|
rows, token = yield self.runInteraction("paginate_file_events", f)
|
||||||
|
|
||||||
|
events = yield self._get_events(
|
||||||
|
[r["event_id"] for r in rows],
|
||||||
|
get_prev_content=True
|
||||||
|
)
|
||||||
|
|
||||||
|
self._set_before_and_after(events, rows)
|
||||||
|
|
||||||
|
defer.returnValue((events, token))
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_recent_events_for_room(self, room_id, limit, end_token, from_token=None):
|
def get_recent_events_for_room(self, room_id, limit, end_token, from_token=None):
|
||||||
rows, token = yield self.get_recent_event_ids_for_room(
|
rows, token = yield self.get_recent_event_ids_for_room(
|
||||||
|
|||||||
@@ -56,7 +56,7 @@ class PaginationConfig(object):
|
|||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def from_request(cls, request, raise_invalid_params=True,
|
def from_request(cls, request, raise_invalid_params=True,
|
||||||
default_limit=None):
|
default_limit=None, default_dir='f'):
|
||||||
def get_param(name, default=None):
|
def get_param(name, default=None):
|
||||||
lst = request.args.get(name, [])
|
lst = request.args.get(name, [])
|
||||||
if len(lst) > 1:
|
if len(lst) > 1:
|
||||||
@@ -68,7 +68,7 @@ class PaginationConfig(object):
|
|||||||
else:
|
else:
|
||||||
return default
|
return default
|
||||||
|
|
||||||
direction = get_param("dir", 'f')
|
direction = get_param("dir", default_dir)
|
||||||
if direction not in ['f', 'b']:
|
if direction not in ['f', 'b']:
|
||||||
raise SynapseError(400, "'dir' parameter is invalid.")
|
raise SynapseError(400, "'dir' parameter is invalid.")
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user