By default, this playbook configures the Synapse Matrix server, so that it works for the general case. If that's okay, you can skip this document.
The playbook provides lots of customization variables you could use to change Synapse's settings.
Their defaults are defined in roles/custom/matrix-synapse/defaults/main.yml
and they ultimately end up in the generated /matrix/synapse/config/homeserver.yaml
file (on the server). This file is generated from the roles/custom/matrix-synapse/templates/synapse/homeserver.yaml.j2
template.
If there's an existing variable which controls a setting you wish to change, you can simply define that variable in your configuration file (inventory/host_vars/matrix.example.com/vars.yml
) and re-run the playbook to apply the changes.
Alternatively, if there is no pre-defined variable for a Synapse setting you wish to change:
-
you can either request a variable to be created (or you can submit such a contribution yourself). Keep in mind that it's probably not a good idea to create variables for each one of Synapse's various settings that rarely get used.
-
or, you can extend and override the default configuration (
homeserver.yaml.j2
) by making use of thematrix_synapse_configuration_extension_yaml
variable. You can find information about this inroles/custom/matrix-synapse/defaults/main.yml
. -
or, if extending the configuration is still not powerful enough for your needs, you can override the configuration completely using
matrix_synapse_configuration
(ormatrix_synapse_configuration_yaml
). You can find information about this inroles/custom/matrix-synapse/defaults/main.yml
.
To have Synapse gracefully handle thousands of users, worker support should be enabled. It factors out some homeserver tasks and spreads the load of incoming client and server-to-server traffic between multiple processes. More information can be found in the official Synapse workers documentation and Tom Foster's Synapse homeserver guide.
To enable Synapse worker support, add the following configuration to your inventory/host_vars/matrix.example.com/vars.yml
file:
matrix_synapse_workers_enabled: true
matrix_synapse_workers_preset: one-of-each
By default, this enables the one-of-each
worker preset, but you may wish to use another preset or control the number of worker instances.
We support a few configuration presets (matrix_synapse_workers_preset: one-of-each
being the default configuration right now):
- (federation-only)
little-federation-helper
- a very minimal worker configuration to improve federation performance - (generic)
one-of-each
- defaults to one worker of each supported type - no smart routing, just generic workers - (specialized)
specialized-workers
- defaults to one worker of each supported type, but disables generic workers and uses specialized workers instead
These presets represent a few common configurations. There are many worker types which can be mixed and matched based on your needs.
Previously, the playbook only supported the most basic type of load-balancing. We call it generic load-balancing below, because incoming HTTP requests are sent to a generic worker. Load-balancing was done based on the requestor's IP address. This is simple, but not necessarily optimal. If you're accessing your account from multiple IP addresses (e.g. your mobile phone being on a different network than your PC), these separate requests may potentially be routed to different workers, each of which would need to cache roughly the same data.
This is still the default load-balancing method (preset) used by the playbook.
To use generic load-balancing, do not specify matrix_synapse_workers_preset
to make it use the default value (one-of-each
), or better yet - explicitly set it as one-of-each
.
You may also consider tweaking the number of workers of each type from the default (one of each).
The playbook now supports a smarter specialized load-balancing inspired by Tom Foster's Synapse homeserver guide. Instead of routing requests to one or more generic workers based only on the requestor's IP adddress, specialized load-balancing routes to 4 different types of specialized workers based on smarter criteria - the access token (username) of the requestor and/or on the resource (room, etc.) being requested.
The playbook supports these 4 types of specialized workers:
- Room workers - handles various Client-Server & Federation APIs dedicated to handling specific rooms
- Sync workers - handles various Client-Server APIs related to synchronization (most notably the
/sync
endpoint) - Client readers - handles various Client-Server APIs which are not for specific rooms (handled by room workers) or for synchronization (handled by sync workers)
- Federation readers - handles various Federation APIs which are not for specific rooms (handled by room workers)
To use specialized load-balancing, consider enabling the specialized-workers
worker preset and potentially tweaking the number of workers of each type from the default (one of each).
If you'd like more customization power, you can start with one of the worker presets and then tweak various matrix_synapse_workers_*_count
variables manually.
To find what variables are available for you to override in your own vars.yml
configuration file, see the defaults/main.yml
file for the matrix-synapse
Ansible role.
The only thing you cannot do is mix generic workers and specialized workers.
When Synapse workers are enabled, the integrated Postgres database is tuned, so that the maximum number of Postgres connections are increased from 200
to 500
. If you need to decrease or increase the number of maximum Postgres connections further, use the postgres_max_connections
variable.
A separate Ansible role (matrix-synapse-reverse-proxy-companion
) and component handles load-balancing for workers. This role/component is automatically enabled when you enable workers. Make sure to use the setup-all
tag (not install-all
!) during the playbook's installation process, especially if you're disabling workers, so that components may be installed/uninstalled correctly.
In case any problems occur, make sure to have a look at the list of synapse issues about workers and your journalctl --unit 'matrix-*'
.
Certain Synapse administration tasks (managing users and rooms, etc.) can be performed via a web user-interace, if you install Synapse Admin.
💡 An alternative to setting up OIDC in Synapse is to use Matrix Authentication Service (MAS). Newer clients (like Element X) only support SSO-based authentication via MAS and not via the legacy Synapse OIDC setup described below. That said, MAS is still a new experimental service which comes with its own downsides. Consult its documentation to learn if it will be a good fit for your deployment.
If you'd like to use OpenID Connect authentication with Synapse, you'll need some additional configuration.
This example configuration is for keycloak, an opensource Identity Provider maintained by Red Hat.
For more detailed documentation on available options and how to setup keycloak, see the Synapse documentation on OpenID Connect with keycloak.
In case you encounter errors regarding the parsing of the variables, you can try to add {% raw %}
and {% endraw %}
blocks around them. For example ;
matrix_synapse_oidc_enabled: true
matrix_synapse_oidc_providers:
- idp_id: keycloak
idp_name: "My KeyCloak server"
issuer: "https://url.ix/auth/realms/{realm_name}"
client_id: "matrix"
client_secret: "{{ vault_synapse_keycloak }}"
scopes: ["openid", "profile"]
user_mapping_provider:
config:
localpart_template: "{% raw %}{{ user.preferred_username }}{% endraw %}"
display_name_template: "{% raw %}{{ user.name }}{% endraw %}"
email_template: "{% raw %}{{ user.email }}{% endraw %}"
allow_existing_users: true # Optional
backchannel_logout_enabled: true # Optional
Templates are used by Synapse for showing certain web pages handled by the server, as well as for email notifications.
This playbook allows you to customize the default templates (see the synapse/res/templates
directory).
If template customization is enabled, the playbook will build a custom container image based on the official one.
Your custom templates need to live in a public or private git repository. This repository will be cloned during Synapse image customization (during the playbook run).
To enable template customizations, use a configuration (inventory/host_vars/matrix.example.com/vars.yml
) like this:
# If you'd like to ensure that the customized image is built each time the playbook runs, enable this.
# Otherwise, the customized image will only be rebuilt whenever the Synapse version changes (once every ~2 weeks).
# matrix_synapse_docker_image_customized_build_nocache: true
matrix_synapse_container_image_customizations_templates_enabled: true
# Our templates live in a templates/ directory within the repository.
# If they're at the root path, delete this line.
matrix_synapse_container_image_customizations_templates_in_container_template_files_relative_path: templates
matrix_synapse_container_image_customizations_templates_git_repository_url: git@github.com:organization/repository.git
matrix_synapse_container_image_customizations_templates_git_repository_branch: main
matrix_synapse_container_image_customizations_templates_git_repository_keyscan_enabled: true
matrix_synapse_container_image_customizations_templates_git_repository_keyscan_hostname: github.com
# If your git repository is public, do not define the private key (remove the variable).
matrix_synapse_container_image_customizations_templates_git_repository_ssh_private_key: |
-----BEGIN OPENSSH PRIVATE KEY-----
....
-----END OPENSSH PRIVATE KEY-----
As mentioned in Synapse's Templates documentation, Synapse will fall back to its own templates if a template is not found in that directory. Due to this, it's recommended to only store and maintain template files in your repository if you need to make custom changes. Other files (which you don't need to change), should not be duplicated, so that you don't need to worry about getting out-of-sync with the original Synapse templates.
This playbook allows you to enable Synapse metrics, which can provide insight into the performance and activity of Synapse.
To enable Synapse runtime metrics see: Enabling metrics and graphs (Prometheus, Grafana) for your Matrix server
To enable Synapse usage metrics, see: Enabling synapse-usage-exporter for Synapse usage statistics