From a3a05c812dbffd143f9a58bdf050bfac9062ba02 Mon Sep 17 00:00:00 2001 From: Erik Johnston Date: Thu, 30 Jun 2022 15:05:49 +0100 Subject: [PATCH 1/2] Add index to help delete old push actions (#13141) --- changelog.d/13141.bugfix | 1 + synapse/_scripts/synapse_port_db.py | 6 ++---- .../databases/main/event_push_actions.py | 10 ++++++++++ .../delta/72/02event_push_actions_index.sql | 19 +++++++++++++++++++ 4 files changed, 32 insertions(+), 4 deletions(-) create mode 100644 changelog.d/13141.bugfix create mode 100644 synapse/storage/schema/main/delta/72/02event_push_actions_index.sql diff --git a/changelog.d/13141.bugfix b/changelog.d/13141.bugfix new file mode 100644 index 000000000000..930e870865e1 --- /dev/null +++ b/changelog.d/13141.bugfix @@ -0,0 +1 @@ +Fix DB performance when deleting old push notifications. Introduced in v1.62.0rc1. diff --git a/synapse/_scripts/synapse_port_db.py b/synapse/_scripts/synapse_port_db.py index 9c06c837dc1d..7b96f61d7b94 100755 --- a/synapse/_scripts/synapse_port_db.py +++ b/synapse/_scripts/synapse_port_db.py @@ -58,9 +58,7 @@ from synapse.storage.databases.main.deviceinbox import DeviceInboxBackgroundUpdateStore from synapse.storage.databases.main.devices import DeviceBackgroundUpdateStore from synapse.storage.databases.main.end_to_end_keys import EndToEndKeyBackgroundStore -from synapse.storage.databases.main.event_push_actions import ( - EventPushActionsWorkerStore, -) +from synapse.storage.databases.main.event_push_actions import EventPushActionsStore from synapse.storage.databases.main.events_bg_updates import ( EventsBackgroundUpdatesStore, ) @@ -202,7 +200,7 @@ class Store( - EventPushActionsWorkerStore, + EventPushActionsStore, ClientIpBackgroundUpdateStore, DeviceInboxBackgroundUpdateStore, DeviceBackgroundUpdateStore, diff --git a/synapse/storage/databases/main/event_push_actions.py b/synapse/storage/databases/main/event_push_actions.py index 80ca2fd0b612..3fadd0930bf3 100644 --- a/synapse/storage/databases/main/event_push_actions.py +++ b/synapse/storage/databases/main/event_push_actions.py @@ -1189,6 +1189,16 @@ def __init__( where_clause="highlight=1", ) + # Add index to make deleting old push actions faster. + self.db_pool.updates.register_background_index_update( + "event_push_actions_stream_highlight_index", + index_name="event_push_actions_stream_highlight_index", + table="event_push_actions", + columns=["highlight", "stream_ordering"], + where_clause="highlight=0", + psql_only=True, + ) + async def get_push_actions_for_user( self, user_id: str, diff --git a/synapse/storage/schema/main/delta/72/02event_push_actions_index.sql b/synapse/storage/schema/main/delta/72/02event_push_actions_index.sql new file mode 100644 index 000000000000..cd0c11d95101 --- /dev/null +++ b/synapse/storage/schema/main/delta/72/02event_push_actions_index.sql @@ -0,0 +1,19 @@ +/* Copyright 2022 The Matrix.org Foundation C.I.C + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +-- Add an index to `event_push_actions` to make deleting old non-highlight push +-- actions faster. +INSERT INTO background_updates (ordering, update_name, progress_json) VALUES + (7202, 'event_push_actions_stream_highlight_index', '{}'); From dbce28b2f14e5a4d0db9f090caf390ea998d1e59 Mon Sep 17 00:00:00 2001 From: Erik Johnston Date: Thu, 30 Jun 2022 15:08:40 +0100 Subject: [PATCH 2/2] Fix unread counts on large servers (#13140) --- changelog.d/13140.bugfix | 1 + .../databases/main/event_push_actions.py | 48 ++++++++++--------- tests/storage/test_event_push_actions.py | 12 ++--- 3 files changed, 32 insertions(+), 29 deletions(-) create mode 100644 changelog.d/13140.bugfix diff --git a/changelog.d/13140.bugfix b/changelog.d/13140.bugfix new file mode 100644 index 000000000000..cb0586e39e88 --- /dev/null +++ b/changelog.d/13140.bugfix @@ -0,0 +1 @@ +Fix unread counts for users on large servers. Introduced in v1.62.0rc1. diff --git a/synapse/storage/databases/main/event_push_actions.py b/synapse/storage/databases/main/event_push_actions.py index 3fadd0930bf3..7d4754b3d37a 100644 --- a/synapse/storage/databases/main/event_push_actions.py +++ b/synapse/storage/databases/main/event_push_actions.py @@ -854,18 +854,20 @@ def _handle_new_receipts_for_notifs_txn(self, txn: LoggingTransaction) -> bool: limit = 100 - min_stream_id = self.db_pool.simple_select_one_onecol_txn( + min_receipts_stream_id = self.db_pool.simple_select_one_onecol_txn( txn, table="event_push_summary_last_receipt_stream_id", keyvalues={}, retcol="stream_id", ) + max_receipts_stream_id = self._receipts_id_gen.get_current_token() + sql = """ SELECT r.stream_id, r.room_id, r.user_id, e.stream_ordering FROM receipts_linearized AS r INNER JOIN events AS e USING (event_id) - WHERE r.stream_id > ? AND user_id LIKE ? + WHERE ? < r.stream_id AND r.stream_id <= ? AND user_id LIKE ? ORDER BY r.stream_id ASC LIMIT ? """ @@ -877,13 +879,21 @@ def _handle_new_receipts_for_notifs_txn(self, txn: LoggingTransaction) -> bool: txn.execute( sql, ( - min_stream_id, + min_receipts_stream_id, + max_receipts_stream_id, user_filter, limit, ), ) rows = txn.fetchall() + old_rotate_stream_ordering = self.db_pool.simple_select_one_onecol_txn( + txn, + table="event_push_summary_stream_ordering", + keyvalues={}, + retcol="stream_ordering", + ) + # For each new read receipt we delete push actions from before it and # recalculate the summary. for _, room_id, user_id, stream_ordering in rows: @@ -902,13 +912,6 @@ def _handle_new_receipts_for_notifs_txn(self, txn: LoggingTransaction) -> bool: (room_id, user_id, stream_ordering), ) - old_rotate_stream_ordering = self.db_pool.simple_select_one_onecol_txn( - txn, - table="event_push_summary_stream_ordering", - keyvalues={}, - retcol="stream_ordering", - ) - notif_count, unread_count = self._get_notif_unread_count_for_user_room( txn, room_id, user_id, stream_ordering, old_rotate_stream_ordering ) @@ -927,18 +930,19 @@ def _handle_new_receipts_for_notifs_txn(self, txn: LoggingTransaction) -> bool: # We always update `event_push_summary_last_receipt_stream_id` to # ensure that we don't rescan the same receipts for remote users. - # - # This requires repeatable read to be safe, as we need the - # `MAX(stream_id)` to not include any new rows that have been committed - # since the start of the transaction (since those rows won't have been - # returned by the query above). Alternatively we could query the max - # stream ID at the start of the transaction and bound everything by - # that. - txn.execute( - """ - UPDATE event_push_summary_last_receipt_stream_id - SET stream_id = (SELECT COALESCE(MAX(stream_id), 0) FROM receipts_linearized) - """ + + upper_limit = max_receipts_stream_id + if len(rows) >= limit: + # If we pulled out a limited number of rows we only update the + # position to the last receipt we processed, so we continue + # processing the rest next iteration. + upper_limit = rows[-1][0] + + self.db_pool.simple_update_txn( + txn, + table="event_push_summary_last_receipt_stream_id", + keyvalues={}, + updatevalues={"stream_id": upper_limit}, ) return len(rows) < limit diff --git a/tests/storage/test_event_push_actions.py b/tests/storage/test_event_push_actions.py index ef069a8110b8..684485ae0629 100644 --- a/tests/storage/test_event_push_actions.py +++ b/tests/storage/test_event_push_actions.py @@ -134,15 +134,12 @@ def _mark_read(stream: int, depth: int) -> None: last_read_stream_ordering[0] = stream self.get_success( - self.store.db_pool.runInteraction( - "", - self.store._insert_linearized_receipt_txn, + self.store.insert_receipt( room_id, "m.read", - user_id, - f"$test{stream}:example.com", - {}, - stream, + user_id=user_id, + event_ids=[f"$test{stream}:example.com"], + data={}, ) ) @@ -166,6 +163,7 @@ def _mark_read(stream: int, depth: int) -> None: _inject_actions(6, PlAIN_NOTIF) _rotate(7) + _assert_counts(1, 0) self.get_success( self.store.db_pool.simple_delete(