mirror of
https://github.com/1f349/dendrite.git
synced 2024-12-29 11:36:28 +00:00
2c581377a5
* Remodel how device list change IDs are created Previously we made them using the offset Kafka supplied. We don't run Kafka anymore, so now we make the SQL table assign the change ID via an AUTOINCREMENTing ID. Redesign the `keyserver_key_changes` table to have `UNIQUE(user_id)` so we don't accumulate key changes forevermore, we now have at most 1 row per user which contains the highest change ID. This needs a SQL migration. * Ensure we bump the change ID on sqlite * Actually read the DeviceChangeID not the Offset in synapi * Add SQL migrations * Prepare after migration; fixup dendrite-upgrade-test logging * Use higher version numbers; fix sqlite query to increment better * Default 0 on postgres * fixup postgres migration on fresh dendrite instances
77 lines
2.2 KiB
Go
77 lines
2.2 KiB
Go
// Copyright 2022 The Matrix.org Foundation C.I.C.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package deltas
|
|
|
|
import (
|
|
"database/sql"
|
|
"fmt"
|
|
|
|
"github.com/matrix-org/dendrite/internal/sqlutil"
|
|
"github.com/pressly/goose"
|
|
)
|
|
|
|
func LoadFromGoose() {
|
|
goose.AddMigration(UpRefactorKeyChanges, DownRefactorKeyChanges)
|
|
}
|
|
|
|
func LoadRefactorKeyChanges(m *sqlutil.Migrations) {
|
|
m.AddMigration(UpRefactorKeyChanges, DownRefactorKeyChanges)
|
|
}
|
|
|
|
func UpRefactorKeyChanges(tx *sql.Tx) error {
|
|
// start counting from the last max offset, else 0.
|
|
var maxOffset int64
|
|
var userID string
|
|
_ = tx.QueryRow(`SELECT user_id, MAX(log_offset) FROM keyserver_key_changes GROUP BY user_id`).Scan(&userID, &maxOffset)
|
|
|
|
_, err := tx.Exec(`
|
|
-- make the new table
|
|
DROP TABLE IF EXISTS keyserver_key_changes;
|
|
CREATE TABLE IF NOT EXISTS keyserver_key_changes (
|
|
change_id INTEGER PRIMARY KEY AUTOINCREMENT,
|
|
-- The key owner
|
|
user_id TEXT NOT NULL,
|
|
UNIQUE (user_id)
|
|
);
|
|
`)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to execute upgrade: %w", err)
|
|
}
|
|
// to start counting from maxOffset, insert a row with that value
|
|
if userID != "" {
|
|
_, err = tx.Exec(`INSERT INTO keyserver_key_changes(change_id, user_id) VALUES($1, $2)`, maxOffset, userID)
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func DownRefactorKeyChanges(tx *sql.Tx) error {
|
|
_, err := tx.Exec(`
|
|
-- Drop all data and revert back, we can't keep the data as Kafka offsets determine the numbers
|
|
DROP TABLE IF EXISTS keyserver_key_changes;
|
|
CREATE TABLE IF NOT EXISTS keyserver_key_changes (
|
|
partition BIGINT NOT NULL,
|
|
offset BIGINT NOT NULL,
|
|
-- The key owner
|
|
user_id TEXT NOT NULL,
|
|
UNIQUE (partition, offset)
|
|
);
|
|
`)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to execute downgrade: %w", err)
|
|
}
|
|
return nil
|
|
}
|