mirror of
https://github.com/1f349/dendrite.git
synced 2024-11-26 05:31:32 +00:00
738686ae68
This adds a new admin endpoint `/_dendrite/admin/purgeRoom/{roomID}`. It completely erases all database entries for a given room ID. The roomserver will start by clearing all data for that room and then will generate an output event to notify downstream components (i.e. the sync API and federation API) to do the same. It does not currently clear media and it is currently not implemented for SQLite since it relies on SQL array operations right now. Co-authored-by: Neil Alexander <neilalexander@users.noreply.github.com> Co-authored-by: Till Faelligen <2353100+S7evinK@users.noreply.github.com>
696 lines
26 KiB
Go
696 lines
26 KiB
Go
// Copyright 2017-2018 New Vector Ltd
|
|
// Copyright 2019-2020 The Matrix.org Foundation C.I.C.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package postgres
|
|
|
|
import (
|
|
"context"
|
|
"database/sql"
|
|
"encoding/json"
|
|
"sort"
|
|
|
|
"github.com/matrix-org/dendrite/internal"
|
|
"github.com/matrix-org/dendrite/roomserver/api"
|
|
"github.com/matrix-org/dendrite/syncapi/storage/postgres/deltas"
|
|
"github.com/matrix-org/dendrite/syncapi/storage/tables"
|
|
"github.com/matrix-org/dendrite/syncapi/types"
|
|
|
|
"github.com/lib/pq"
|
|
"github.com/matrix-org/gomatrixserverlib"
|
|
|
|
"github.com/matrix-org/dendrite/internal/sqlutil"
|
|
)
|
|
|
|
const outputRoomEventsSchema = `
|
|
-- This sequence is shared between all the tables generated from kafka logs.
|
|
CREATE SEQUENCE IF NOT EXISTS syncapi_stream_id;
|
|
|
|
-- Stores output room events received from the roomserver.
|
|
CREATE TABLE IF NOT EXISTS syncapi_output_room_events (
|
|
-- An incrementing ID which denotes the position in the log that this event resides at.
|
|
-- NB: 'serial' makes no guarantees to increment by 1 every time, only that it increments.
|
|
-- This isn't a problem for us since we just want to order by this field.
|
|
id BIGINT PRIMARY KEY DEFAULT nextval('syncapi_stream_id'),
|
|
-- The event ID for the event
|
|
event_id TEXT NOT NULL CONSTRAINT syncapi_event_id_idx UNIQUE,
|
|
-- The 'room_id' key for the event.
|
|
room_id TEXT NOT NULL,
|
|
-- The headered JSON for the event, containing potentially additional metadata such as
|
|
-- the room version. Stored as TEXT because this should be valid UTF-8.
|
|
headered_event_json TEXT NOT NULL,
|
|
-- The event type e.g 'm.room.member'.
|
|
type TEXT NOT NULL,
|
|
-- The 'sender' property of the event.
|
|
sender TEXT NOT NULL,
|
|
-- true if the event content contains a url key.
|
|
contains_url BOOL NOT NULL,
|
|
-- A list of event IDs which represent a delta of added/removed room state. This can be NULL
|
|
-- if there is no delta.
|
|
add_state_ids TEXT[],
|
|
remove_state_ids TEXT[],
|
|
-- The client session that sent the event, if any
|
|
session_id BIGINT,
|
|
-- The transaction id used to send the event, if any
|
|
transaction_id TEXT,
|
|
-- Should the event be excluded from responses to /sync requests. Useful for
|
|
-- events retrieved through backfilling that have a position in the stream
|
|
-- that relates to the moment these were retrieved rather than the moment these
|
|
-- were emitted.
|
|
exclude_from_sync BOOL DEFAULT FALSE,
|
|
-- The history visibility before this event (1 - world_readable; 2 - shared; 3 - invited; 4 - joined)
|
|
history_visibility SMALLINT NOT NULL DEFAULT 2
|
|
);
|
|
|
|
CREATE INDEX IF NOT EXISTS syncapi_output_room_events_type_idx ON syncapi_output_room_events (type);
|
|
CREATE INDEX IF NOT EXISTS syncapi_output_room_events_sender_idx ON syncapi_output_room_events (sender);
|
|
CREATE INDEX IF NOT EXISTS syncapi_output_room_events_room_id_idx ON syncapi_output_room_events (room_id);
|
|
CREATE INDEX IF NOT EXISTS syncapi_output_room_events_exclude_from_sync_idx ON syncapi_output_room_events (exclude_from_sync);
|
|
CREATE INDEX IF NOT EXISTS syncapi_output_room_events_add_state_ids_idx ON syncapi_output_room_events ((add_state_ids IS NOT NULL));
|
|
CREATE INDEX IF NOT EXISTS syncapi_output_room_events_remove_state_ids_idx ON syncapi_output_room_events ((remove_state_ids IS NOT NULL));
|
|
`
|
|
|
|
const insertEventSQL = "" +
|
|
"INSERT INTO syncapi_output_room_events (" +
|
|
"room_id, event_id, headered_event_json, type, sender, contains_url, add_state_ids, remove_state_ids, session_id, transaction_id, exclude_from_sync, history_visibility" +
|
|
") VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11, $12) " +
|
|
"ON CONFLICT ON CONSTRAINT syncapi_event_id_idx DO UPDATE SET exclude_from_sync = (excluded.exclude_from_sync AND $11) " +
|
|
"RETURNING id"
|
|
|
|
const selectEventsSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id, history_visibility FROM syncapi_output_room_events WHERE event_id = ANY($1)"
|
|
|
|
const selectEventsWithFilterSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id, history_visibility FROM syncapi_output_room_events WHERE event_id = ANY($1)" +
|
|
" AND ( $2::text[] IS NULL OR sender = ANY($2) )" +
|
|
" AND ( $3::text[] IS NULL OR NOT(sender = ANY($3)) )" +
|
|
" AND ( $4::text[] IS NULL OR type LIKE ANY($4) )" +
|
|
" AND ( $5::text[] IS NULL OR NOT(type LIKE ANY($5)) )" +
|
|
" AND ( $6::bool IS NULL OR contains_url = $6 )" +
|
|
" LIMIT $7"
|
|
|
|
const selectRecentEventsSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id, history_visibility FROM syncapi_output_room_events" +
|
|
" WHERE room_id = $1 AND id > $2 AND id <= $3" +
|
|
" AND ( $4::text[] IS NULL OR sender = ANY($4) )" +
|
|
" AND ( $5::text[] IS NULL OR NOT(sender = ANY($5)) )" +
|
|
" AND ( $6::text[] IS NULL OR type LIKE ANY($6) )" +
|
|
" AND ( $7::text[] IS NULL OR NOT(type LIKE ANY($7)) )" +
|
|
" ORDER BY id DESC LIMIT $8"
|
|
|
|
const selectRecentEventsForSyncSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id, history_visibility FROM syncapi_output_room_events" +
|
|
" WHERE room_id = $1 AND id > $2 AND id <= $3 AND exclude_from_sync = FALSE" +
|
|
" AND ( $4::text[] IS NULL OR sender = ANY($4) )" +
|
|
" AND ( $5::text[] IS NULL OR NOT(sender = ANY($5)) )" +
|
|
" AND ( $6::text[] IS NULL OR type LIKE ANY($6) )" +
|
|
" AND ( $7::text[] IS NULL OR NOT(type LIKE ANY($7)) )" +
|
|
" ORDER BY id DESC LIMIT $8"
|
|
|
|
const selectEarlyEventsSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id, history_visibility FROM syncapi_output_room_events" +
|
|
" WHERE room_id = $1 AND id > $2 AND id <= $3" +
|
|
" AND ( $4::text[] IS NULL OR sender = ANY($4) )" +
|
|
" AND ( $5::text[] IS NULL OR NOT(sender = ANY($5)) )" +
|
|
" AND ( $6::text[] IS NULL OR type LIKE ANY($6) )" +
|
|
" AND ( $7::text[] IS NULL OR NOT(type LIKE ANY($7)) )" +
|
|
" ORDER BY id ASC LIMIT $8"
|
|
|
|
const selectMaxEventIDSQL = "" +
|
|
"SELECT MAX(id) FROM syncapi_output_room_events"
|
|
|
|
const updateEventJSONSQL = "" +
|
|
"UPDATE syncapi_output_room_events SET headered_event_json=$1 WHERE event_id=$2"
|
|
|
|
// In order for us to apply the state updates correctly, rows need to be ordered in the order they were received (id).
|
|
const selectStateInRangeFilteredSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, exclude_from_sync, add_state_ids, remove_state_ids, history_visibility" +
|
|
" FROM syncapi_output_room_events" +
|
|
" WHERE (id > $1 AND id <= $2) AND (add_state_ids IS NOT NULL OR remove_state_ids IS NOT NULL)" +
|
|
" AND room_id = ANY($3)" +
|
|
" AND ( $4::text[] IS NULL OR sender = ANY($4) )" +
|
|
" AND ( $5::text[] IS NULL OR NOT(sender = ANY($5)) )" +
|
|
" AND ( $6::text[] IS NULL OR type LIKE ANY($6) )" +
|
|
" AND ( $7::text[] IS NULL OR NOT(type LIKE ANY($7)) )" +
|
|
" AND ( $8::bool IS NULL OR contains_url = $8 )" +
|
|
" ORDER BY id ASC"
|
|
|
|
// In order for us to apply the state updates correctly, rows need to be ordered in the order they were received (id).
|
|
const selectStateInRangeSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, exclude_from_sync, add_state_ids, remove_state_ids, history_visibility" +
|
|
" FROM syncapi_output_room_events" +
|
|
" WHERE (id > $1 AND id <= $2) AND (add_state_ids IS NOT NULL OR remove_state_ids IS NOT NULL)" +
|
|
" AND room_id = ANY($3)" +
|
|
" ORDER BY id ASC"
|
|
|
|
const deleteEventsForRoomSQL = "" +
|
|
"DELETE FROM syncapi_output_room_events WHERE room_id = $1"
|
|
|
|
const selectContextEventSQL = "" +
|
|
"SELECT id, headered_event_json, history_visibility FROM syncapi_output_room_events WHERE room_id = $1 AND event_id = $2"
|
|
|
|
const selectContextBeforeEventSQL = "" +
|
|
"SELECT headered_event_json, history_visibility FROM syncapi_output_room_events WHERE room_id = $1 AND id < $2" +
|
|
" AND ( $4::text[] IS NULL OR sender = ANY($4) )" +
|
|
" AND ( $5::text[] IS NULL OR NOT(sender = ANY($5)) )" +
|
|
" AND ( $6::text[] IS NULL OR type LIKE ANY($6) )" +
|
|
" AND ( $7::text[] IS NULL OR NOT(type LIKE ANY($7)) )" +
|
|
" ORDER BY id DESC LIMIT $3"
|
|
|
|
const selectContextAfterEventSQL = "" +
|
|
"SELECT id, headered_event_json, history_visibility FROM syncapi_output_room_events WHERE room_id = $1 AND id > $2" +
|
|
" AND ( $4::text[] IS NULL OR sender = ANY($4) )" +
|
|
" AND ( $5::text[] IS NULL OR NOT(sender = ANY($5)) )" +
|
|
" AND ( $6::text[] IS NULL OR type LIKE ANY($6) )" +
|
|
" AND ( $7::text[] IS NULL OR NOT(type LIKE ANY($7)) )" +
|
|
" ORDER BY id ASC LIMIT $3"
|
|
|
|
const purgeEventsSQL = "" +
|
|
"DELETE FROM syncapi_output_room_events WHERE room_id = $1"
|
|
|
|
const selectSearchSQL = "SELECT id, event_id, headered_event_json FROM syncapi_output_room_events WHERE id > $1 AND type = ANY($2) ORDER BY id ASC LIMIT $3"
|
|
|
|
type outputRoomEventsStatements struct {
|
|
insertEventStmt *sql.Stmt
|
|
selectEventsStmt *sql.Stmt
|
|
selectEventsWitFilterStmt *sql.Stmt
|
|
selectMaxEventIDStmt *sql.Stmt
|
|
selectRecentEventsStmt *sql.Stmt
|
|
selectRecentEventsForSyncStmt *sql.Stmt
|
|
selectEarlyEventsStmt *sql.Stmt
|
|
selectStateInRangeFilteredStmt *sql.Stmt
|
|
selectStateInRangeStmt *sql.Stmt
|
|
updateEventJSONStmt *sql.Stmt
|
|
deleteEventsForRoomStmt *sql.Stmt
|
|
selectContextEventStmt *sql.Stmt
|
|
selectContextBeforeEventStmt *sql.Stmt
|
|
selectContextAfterEventStmt *sql.Stmt
|
|
purgeEventsStmt *sql.Stmt
|
|
selectSearchStmt *sql.Stmt
|
|
}
|
|
|
|
func NewPostgresEventsTable(db *sql.DB) (tables.Events, error) {
|
|
s := &outputRoomEventsStatements{}
|
|
_, err := db.Exec(outputRoomEventsSchema)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
m := sqlutil.NewMigrator(db)
|
|
m.AddMigrations(
|
|
sqlutil.Migration{
|
|
Version: "syncapi: add history visibility column (output_room_events)",
|
|
Up: deltas.UpAddHistoryVisibilityColumnOutputRoomEvents,
|
|
},
|
|
)
|
|
err = m.Up(context.Background())
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return s, sqlutil.StatementList{
|
|
{&s.insertEventStmt, insertEventSQL},
|
|
{&s.selectEventsStmt, selectEventsSQL},
|
|
{&s.selectEventsWitFilterStmt, selectEventsWithFilterSQL},
|
|
{&s.selectMaxEventIDStmt, selectMaxEventIDSQL},
|
|
{&s.selectRecentEventsStmt, selectRecentEventsSQL},
|
|
{&s.selectRecentEventsForSyncStmt, selectRecentEventsForSyncSQL},
|
|
{&s.selectEarlyEventsStmt, selectEarlyEventsSQL},
|
|
{&s.selectStateInRangeFilteredStmt, selectStateInRangeFilteredSQL},
|
|
{&s.selectStateInRangeStmt, selectStateInRangeSQL},
|
|
{&s.updateEventJSONStmt, updateEventJSONSQL},
|
|
{&s.deleteEventsForRoomStmt, deleteEventsForRoomSQL},
|
|
{&s.selectContextEventStmt, selectContextEventSQL},
|
|
{&s.selectContextBeforeEventStmt, selectContextBeforeEventSQL},
|
|
{&s.selectContextAfterEventStmt, selectContextAfterEventSQL},
|
|
{&s.purgeEventsStmt, purgeEventsSQL},
|
|
{&s.selectSearchStmt, selectSearchSQL},
|
|
}.Prepare(db)
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) UpdateEventJSON(ctx context.Context, txn *sql.Tx, event *gomatrixserverlib.HeaderedEvent) error {
|
|
headeredJSON, err := json.Marshal(event)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
_, err = sqlutil.TxStmt(txn, s.updateEventJSONStmt).ExecContext(ctx, headeredJSON, event.EventID())
|
|
return err
|
|
}
|
|
|
|
// selectStateInRange returns the state events between the two given PDU stream positions, exclusive of oldPos, inclusive of newPos.
|
|
// Results are bucketed based on the room ID. If the same state is overwritten multiple times between the
|
|
// two positions, only the most recent state is returned.
|
|
func (s *outputRoomEventsStatements) SelectStateInRange(
|
|
ctx context.Context, txn *sql.Tx, r types.Range,
|
|
stateFilter *gomatrixserverlib.StateFilter, roomIDs []string,
|
|
) (map[string]map[string]bool, map[string]types.StreamEvent, error) {
|
|
var rows *sql.Rows
|
|
var err error
|
|
if stateFilter != nil {
|
|
stmt := sqlutil.TxStmt(txn, s.selectStateInRangeFilteredStmt)
|
|
senders, notSenders := getSendersStateFilterFilter(stateFilter)
|
|
rows, err = stmt.QueryContext(
|
|
ctx, r.Low(), r.High(), pq.StringArray(roomIDs),
|
|
pq.StringArray(senders),
|
|
pq.StringArray(notSenders),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(stateFilter.Types)),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(stateFilter.NotTypes)),
|
|
stateFilter.ContainsURL,
|
|
)
|
|
} else {
|
|
stmt := sqlutil.TxStmt(txn, s.selectStateInRangeStmt)
|
|
rows, err = stmt.QueryContext(
|
|
ctx, r.Low(), r.High(), pq.StringArray(roomIDs),
|
|
)
|
|
}
|
|
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "selectStateInRange: rows.close() failed")
|
|
// Fetch all the state change events for all rooms between the two positions then loop each event and:
|
|
// - Keep a cache of the event by ID (99% of state change events are for the event itself)
|
|
// - For each room ID, build up an array of event IDs which represents cumulative adds/removes
|
|
// For each room, map cumulative event IDs to events and return. This may need to a batch SELECT based on event ID
|
|
// if they aren't in the event ID cache. We don't handle state deletion yet.
|
|
eventIDToEvent := make(map[string]types.StreamEvent)
|
|
|
|
// RoomID => A set (map[string]bool) of state event IDs which are between the two positions
|
|
stateNeeded := make(map[string]map[string]bool)
|
|
|
|
for rows.Next() {
|
|
var (
|
|
eventID string
|
|
streamPos types.StreamPosition
|
|
eventBytes []byte
|
|
excludeFromSync bool
|
|
addIDs pq.StringArray
|
|
delIDs pq.StringArray
|
|
historyVisibility gomatrixserverlib.HistoryVisibility
|
|
)
|
|
if err := rows.Scan(&eventID, &streamPos, &eventBytes, &excludeFromSync, &addIDs, &delIDs, &historyVisibility); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
// TODO: Handle redacted events
|
|
var ev gomatrixserverlib.HeaderedEvent
|
|
if err := ev.UnmarshalJSONWithEventID(eventBytes, eventID); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
needSet := stateNeeded[ev.RoomID()]
|
|
if needSet == nil { // make set if required
|
|
needSet = make(map[string]bool)
|
|
}
|
|
for _, id := range delIDs {
|
|
needSet[id] = false
|
|
}
|
|
for _, id := range addIDs {
|
|
needSet[id] = true
|
|
}
|
|
stateNeeded[ev.RoomID()] = needSet
|
|
ev.Visibility = historyVisibility
|
|
|
|
eventIDToEvent[eventID] = types.StreamEvent{
|
|
HeaderedEvent: &ev,
|
|
StreamPosition: streamPos,
|
|
ExcludeFromSync: excludeFromSync,
|
|
}
|
|
}
|
|
|
|
return stateNeeded, eventIDToEvent, rows.Err()
|
|
}
|
|
|
|
// MaxID returns the ID of the last inserted event in this table. 'txn' is optional. If it is not supplied,
|
|
// then this function should only ever be used at startup, as it will race with inserting events if it is
|
|
// done afterwards. If there are no inserted events, 0 is returned.
|
|
func (s *outputRoomEventsStatements) SelectMaxEventID(
|
|
ctx context.Context, txn *sql.Tx,
|
|
) (id int64, err error) {
|
|
var nullableID sql.NullInt64
|
|
stmt := sqlutil.TxStmt(txn, s.selectMaxEventIDStmt)
|
|
err = stmt.QueryRowContext(ctx).Scan(&nullableID)
|
|
if nullableID.Valid {
|
|
id = nullableID.Int64
|
|
}
|
|
return
|
|
}
|
|
|
|
// InsertEvent into the output_room_events table. addState and removeState are an optional list of state event IDs. Returns the position
|
|
// of the inserted event.
|
|
func (s *outputRoomEventsStatements) InsertEvent(
|
|
ctx context.Context, txn *sql.Tx,
|
|
event *gomatrixserverlib.HeaderedEvent, addState, removeState []string,
|
|
transactionID *api.TransactionID, excludeFromSync bool, historyVisibility gomatrixserverlib.HistoryVisibility,
|
|
) (streamPos types.StreamPosition, err error) {
|
|
var txnID *string
|
|
var sessionID *int64
|
|
if transactionID != nil {
|
|
sessionID = &transactionID.SessionID
|
|
txnID = &transactionID.TransactionID
|
|
}
|
|
|
|
// Parse content as JSON and search for an "url" key
|
|
containsURL := false
|
|
var content map[string]interface{}
|
|
if json.Unmarshal(event.Content(), &content) == nil {
|
|
// Set containsURL to true if url is present
|
|
_, containsURL = content["url"]
|
|
}
|
|
|
|
var headeredJSON []byte
|
|
headeredJSON, err = json.Marshal(event)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
stmt := sqlutil.TxStmt(txn, s.insertEventStmt)
|
|
err = stmt.QueryRowContext(
|
|
ctx,
|
|
event.RoomID(),
|
|
event.EventID(),
|
|
headeredJSON,
|
|
event.Type(),
|
|
event.Sender(),
|
|
containsURL,
|
|
pq.StringArray(addState),
|
|
pq.StringArray(removeState),
|
|
sessionID,
|
|
txnID,
|
|
excludeFromSync,
|
|
historyVisibility,
|
|
).Scan(&streamPos)
|
|
return
|
|
}
|
|
|
|
// selectRecentEvents returns the most recent events in the given room, up to a maximum of 'limit'.
|
|
// If onlySyncEvents has a value of true, only returns the events that aren't marked as to exclude
|
|
// from sync.
|
|
func (s *outputRoomEventsStatements) SelectRecentEvents(
|
|
ctx context.Context, txn *sql.Tx,
|
|
roomID string, r types.Range, eventFilter *gomatrixserverlib.RoomEventFilter,
|
|
chronologicalOrder bool, onlySyncEvents bool,
|
|
) ([]types.StreamEvent, bool, error) {
|
|
var stmt *sql.Stmt
|
|
if onlySyncEvents {
|
|
stmt = sqlutil.TxStmt(txn, s.selectRecentEventsForSyncStmt)
|
|
} else {
|
|
stmt = sqlutil.TxStmt(txn, s.selectRecentEventsStmt)
|
|
}
|
|
senders, notSenders := getSendersRoomEventFilter(eventFilter)
|
|
rows, err := stmt.QueryContext(
|
|
ctx, roomID, r.Low(), r.High(),
|
|
pq.StringArray(senders),
|
|
pq.StringArray(notSenders),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(eventFilter.Types)),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(eventFilter.NotTypes)),
|
|
eventFilter.Limit+1,
|
|
)
|
|
if err != nil {
|
|
return nil, false, err
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "selectRecentEvents: rows.close() failed")
|
|
events, err := rowsToStreamEvents(rows)
|
|
if err != nil {
|
|
return nil, false, err
|
|
}
|
|
if chronologicalOrder {
|
|
// The events need to be returned from oldest to latest, which isn't
|
|
// necessary the way the SQL query returns them, so a sort is necessary to
|
|
// ensure the events are in the right order in the slice.
|
|
sort.SliceStable(events, func(i int, j int) bool {
|
|
return events[i].StreamPosition < events[j].StreamPosition
|
|
})
|
|
}
|
|
// we queried for 1 more than the limit, so if we returned one more mark limited=true
|
|
limited := false
|
|
if len(events) > eventFilter.Limit {
|
|
limited = true
|
|
// re-slice the extra (oldest) event out: in chronological order this is the first entry, else the last.
|
|
if chronologicalOrder {
|
|
events = events[1:]
|
|
} else {
|
|
events = events[:len(events)-1]
|
|
}
|
|
}
|
|
|
|
return events, limited, nil
|
|
}
|
|
|
|
// selectEarlyEvents returns the earliest events in the given room, starting
|
|
// from a given position, up to a maximum of 'limit'.
|
|
func (s *outputRoomEventsStatements) SelectEarlyEvents(
|
|
ctx context.Context, txn *sql.Tx,
|
|
roomID string, r types.Range, eventFilter *gomatrixserverlib.RoomEventFilter,
|
|
) ([]types.StreamEvent, error) {
|
|
senders, notSenders := getSendersRoomEventFilter(eventFilter)
|
|
stmt := sqlutil.TxStmt(txn, s.selectEarlyEventsStmt)
|
|
rows, err := stmt.QueryContext(
|
|
ctx, roomID, r.Low(), r.High(),
|
|
pq.StringArray(senders),
|
|
pq.StringArray(notSenders),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(eventFilter.Types)),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(eventFilter.NotTypes)),
|
|
eventFilter.Limit,
|
|
)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "selectEarlyEvents: rows.close() failed")
|
|
events, err := rowsToStreamEvents(rows)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
// The events need to be returned from oldest to latest, which isn't
|
|
// necessarily the way the SQL query returns them, so a sort is necessary to
|
|
// ensure the events are in the right order in the slice.
|
|
sort.SliceStable(events, func(i int, j int) bool {
|
|
return events[i].StreamPosition < events[j].StreamPosition
|
|
})
|
|
return events, nil
|
|
}
|
|
|
|
// selectEvents returns the events for the given event IDs. If an event is
|
|
// missing from the database, it will be omitted.
|
|
func (s *outputRoomEventsStatements) SelectEvents(
|
|
ctx context.Context, txn *sql.Tx, eventIDs []string, filter *gomatrixserverlib.RoomEventFilter, preserveOrder bool,
|
|
) ([]types.StreamEvent, error) {
|
|
var (
|
|
stmt *sql.Stmt
|
|
rows *sql.Rows
|
|
err error
|
|
)
|
|
if filter == nil {
|
|
stmt = sqlutil.TxStmt(txn, s.selectEventsStmt)
|
|
rows, err = stmt.QueryContext(ctx, pq.StringArray(eventIDs))
|
|
} else {
|
|
senders, notSenders := getSendersRoomEventFilter(filter)
|
|
stmt = sqlutil.TxStmt(txn, s.selectEventsWitFilterStmt)
|
|
rows, err = stmt.QueryContext(ctx,
|
|
pq.StringArray(eventIDs),
|
|
pq.StringArray(senders),
|
|
pq.StringArray(notSenders),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(filter.Types)),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(filter.NotTypes)),
|
|
filter.ContainsURL,
|
|
filter.Limit,
|
|
)
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "selectEvents: rows.close() failed")
|
|
streamEvents, err := rowsToStreamEvents(rows)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if preserveOrder {
|
|
eventMap := make(map[string]types.StreamEvent)
|
|
for _, ev := range streamEvents {
|
|
eventMap[ev.EventID()] = ev
|
|
}
|
|
var returnEvents []types.StreamEvent
|
|
for _, eventID := range eventIDs {
|
|
ev, ok := eventMap[eventID]
|
|
if ok {
|
|
returnEvents = append(returnEvents, ev)
|
|
}
|
|
}
|
|
return returnEvents, nil
|
|
}
|
|
return streamEvents, nil
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) DeleteEventsForRoom(
|
|
ctx context.Context, txn *sql.Tx, roomID string,
|
|
) (err error) {
|
|
_, err = sqlutil.TxStmt(txn, s.deleteEventsForRoomStmt).ExecContext(ctx, roomID)
|
|
return err
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) SelectContextEvent(ctx context.Context, txn *sql.Tx, roomID, eventID string) (id int, evt gomatrixserverlib.HeaderedEvent, err error) {
|
|
row := sqlutil.TxStmt(txn, s.selectContextEventStmt).QueryRowContext(ctx, roomID, eventID)
|
|
|
|
var eventAsString string
|
|
var historyVisibility gomatrixserverlib.HistoryVisibility
|
|
if err = row.Scan(&id, &eventAsString, &historyVisibility); err != nil {
|
|
return 0, evt, err
|
|
}
|
|
|
|
if err = json.Unmarshal([]byte(eventAsString), &evt); err != nil {
|
|
return 0, evt, err
|
|
}
|
|
evt.Visibility = historyVisibility
|
|
return id, evt, nil
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) SelectContextBeforeEvent(
|
|
ctx context.Context, txn *sql.Tx, id int, roomID string, filter *gomatrixserverlib.RoomEventFilter,
|
|
) (evts []*gomatrixserverlib.HeaderedEvent, err error) {
|
|
senders, notSenders := getSendersRoomEventFilter(filter)
|
|
rows, err := sqlutil.TxStmt(txn, s.selectContextBeforeEventStmt).QueryContext(
|
|
ctx, roomID, id, filter.Limit,
|
|
pq.StringArray(senders),
|
|
pq.StringArray(notSenders),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(filter.Types)),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(filter.NotTypes)),
|
|
)
|
|
if err != nil {
|
|
return
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "rows.close() failed")
|
|
|
|
for rows.Next() {
|
|
var (
|
|
eventBytes []byte
|
|
evt *gomatrixserverlib.HeaderedEvent
|
|
historyVisibility gomatrixserverlib.HistoryVisibility
|
|
)
|
|
if err = rows.Scan(&eventBytes, &historyVisibility); err != nil {
|
|
return evts, err
|
|
}
|
|
if err = json.Unmarshal(eventBytes, &evt); err != nil {
|
|
return evts, err
|
|
}
|
|
evt.Visibility = historyVisibility
|
|
evts = append(evts, evt)
|
|
}
|
|
|
|
return evts, rows.Err()
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) SelectContextAfterEvent(
|
|
ctx context.Context, txn *sql.Tx, id int, roomID string, filter *gomatrixserverlib.RoomEventFilter,
|
|
) (lastID int, evts []*gomatrixserverlib.HeaderedEvent, err error) {
|
|
senders, notSenders := getSendersRoomEventFilter(filter)
|
|
rows, err := sqlutil.TxStmt(txn, s.selectContextAfterEventStmt).QueryContext(
|
|
ctx, roomID, id, filter.Limit,
|
|
pq.StringArray(senders),
|
|
pq.StringArray(notSenders),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(filter.Types)),
|
|
pq.StringArray(filterConvertTypeWildcardToSQL(filter.NotTypes)),
|
|
)
|
|
if err != nil {
|
|
return
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "rows.close() failed")
|
|
|
|
for rows.Next() {
|
|
var (
|
|
eventBytes []byte
|
|
evt *gomatrixserverlib.HeaderedEvent
|
|
historyVisibility gomatrixserverlib.HistoryVisibility
|
|
)
|
|
if err = rows.Scan(&lastID, &eventBytes, &historyVisibility); err != nil {
|
|
return 0, evts, err
|
|
}
|
|
if err = json.Unmarshal(eventBytes, &evt); err != nil {
|
|
return 0, evts, err
|
|
}
|
|
evt.Visibility = historyVisibility
|
|
evts = append(evts, evt)
|
|
}
|
|
|
|
return lastID, evts, rows.Err()
|
|
}
|
|
|
|
func rowsToStreamEvents(rows *sql.Rows) ([]types.StreamEvent, error) {
|
|
var result []types.StreamEvent
|
|
for rows.Next() {
|
|
var (
|
|
eventID string
|
|
streamPos types.StreamPosition
|
|
eventBytes []byte
|
|
excludeFromSync bool
|
|
sessionID *int64
|
|
txnID *string
|
|
transactionID *api.TransactionID
|
|
historyVisibility gomatrixserverlib.HistoryVisibility
|
|
)
|
|
if err := rows.Scan(&eventID, &streamPos, &eventBytes, &sessionID, &excludeFromSync, &txnID, &historyVisibility); err != nil {
|
|
return nil, err
|
|
}
|
|
// TODO: Handle redacted events
|
|
var ev gomatrixserverlib.HeaderedEvent
|
|
if err := ev.UnmarshalJSONWithEventID(eventBytes, eventID); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if sessionID != nil && txnID != nil {
|
|
transactionID = &api.TransactionID{
|
|
SessionID: *sessionID,
|
|
TransactionID: *txnID,
|
|
}
|
|
}
|
|
ev.Visibility = historyVisibility
|
|
result = append(result, types.StreamEvent{
|
|
HeaderedEvent: &ev,
|
|
StreamPosition: streamPos,
|
|
TransactionID: transactionID,
|
|
ExcludeFromSync: excludeFromSync,
|
|
})
|
|
}
|
|
return result, rows.Err()
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) PurgeEvents(
|
|
ctx context.Context, txn *sql.Tx, roomID string,
|
|
) error {
|
|
_, err := sqlutil.TxStmt(txn, s.purgeEventsStmt).ExecContext(ctx, roomID)
|
|
return err
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) ReIndex(ctx context.Context, txn *sql.Tx, limit, afterID int64, types []string) (map[int64]gomatrixserverlib.HeaderedEvent, error) {
|
|
rows, err := sqlutil.TxStmt(txn, s.selectSearchStmt).QueryContext(ctx, afterID, pq.StringArray(types), limit)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "rows.close() failed")
|
|
|
|
var eventID string
|
|
var id int64
|
|
result := make(map[int64]gomatrixserverlib.HeaderedEvent)
|
|
for rows.Next() {
|
|
var ev gomatrixserverlib.HeaderedEvent
|
|
var eventBytes []byte
|
|
if err = rows.Scan(&id, &eventID, &eventBytes); err != nil {
|
|
return nil, err
|
|
}
|
|
if err = ev.UnmarshalJSONWithEventID(eventBytes, eventID); err != nil {
|
|
return nil, err
|
|
}
|
|
result[id] = ev
|
|
}
|
|
return result, rows.Err()
|
|
}
|