matrix-media-repo is a highly customizable multi-domain media repository for Matrix. Intended for medium to large environments consisting of several homeservers, this media repo de-duplicates media (including remote media) while being fully compliant with the specification.
Smaller/individual homeservers can still make use of this project's features, though it may be difficult to set up or have higher than expected resource consumption - please do your research before deploying this as this project may not be useful for your environment.
For installation instructions, see docs.t2bot.io.
For deployment information, see docs.t2bot.io.
To properly run the media repo in a development setting, it must be compiled manually once to ensure the assets are set up correctly: follow the compilation steps posted on docs.t2bot.io.
If you'd like to use a regular Matrix client to test the media repo,
docker-compose -f dev/docker-compose.yaml up
will give you a Conduit homeserver behind an nginx reverse proxy which routes media requests to
http://host.docker.internal:8001. To test accurately, it is recommended to add the following homeserver configuration
to your media repo config:
name: "localhost" csApi: "http://localhost:8008" # This is exposed by the nginx container
Federated media requests should function normally with this setup, though the homeserver itself will be unable to federate. For convenience, an element-web instance is also hosted at the same port from the root.
A postgresql server is also created by the docker stack for ease of use. To use it, add the following to your configuration:
database: postgres: "postgres://postgres:firstname.lastname@example.org:5432/postgres?sslmode=disable" pool: maxConnections: 10 maxIdleConnections: 10
Note that the postgresql image is insecure and not recommended for production use. It also does not follow best practices for database management - use at your own risk.
Importing media from synapse
Media is imported by connecting to your synapse database and downloading all the content from the homeserver. This is so you have a backup of the media repository still with synapse. Do not point traffic at the media repo until after the import is complete.
Note: the database options provided on the command line are for the Synapse database. The media repo will use the connection string in the media-repo.yaml config when trying to store the Synapse media.
Note: the import script is not available to the Docker container. Binaries of the script are included with every release though if you want to avoid building it yourself.
- Build the media repo (as stated above)
media-repo.yamlper the install instructions above
bin/import_synapse. The usage is below.
Assuming the media repository, postgres database, and synapse are all on the same host, the command to run would look something like:
Usage of import_synapse: -baseUrl string The base URL to access your homeserver with (default "http://localhost:8008") -config string The path to the media repo configuration (with the database section completed) (default "media-repo.yaml") -dbHost string The PostgresSQL hostname for your Synapse database (default "localhost") -dbName string The name of your Synapse database (default "synapse") -dbPassword string The password for your Synapse's PostgreSQL database. Can be omitted to be prompted when run -dbPort int The port for your Synapse's PostgreSQL database (default 5432) -dbUsername string The username for your Synapse's PostgreSQL database (default "synapse") -migrations string The absolute path the media repo's migrations folder (default "./migrations") -serverName string The name of your homeserver (eg: matrix.org) (default "localhost") -workers int The number of workers to use when downloading media. Using multiple workers risks deduplication not working as efficiently. (default 1)
bin/import_synapse -serverName myserver.com -dbUsername my_database_user -dbName synapse
- Wait for the import to complete. The script will automatically deduplicate media.
- Point traffic to the media repository.
Export and import user data
The admin API for this is specified in docs/admin.md, though they can be difficult to use for scripts.
bin/gdpr_import binaries do the process for you, and do so in memory but against the real
media repo database and datastores - this moves the resource intensiveness to the binary you're running instead of the
media repo instance, but still makes reads and writes to your database and datastores. For example, when exporting a
user's data the binary will pull all the data locally and write it to disk for you, but during that process the user's
export is accessible via the main media repo too. The export is deleted if the binary is successful at exporting the
Note: Imports done through this method can affect other homeservers! For example, a user's data export could contain an entry for a homeserver other than their own, which the media repo will happily import. Always validate the manifest of an import before running it!
Ensuring you have your media repo config available, here's the help for each binary:
Usage of gdpr_export: -config string The path to the configuration (default "media-repo.yaml") -destination string The directory for where export files should be placed (default "./gdpr-data") -entity string The user ID or server name to export -migrations string The absolute path for the migrations folder (default "./migrations") -templates string The absolute path for the templates folder (default "./templates")
Usage of gdpr_import: -config string The path to the configuration (default "media-repo.yaml") -directory string The directory for where the entity's exported files are (default "./gdpr-data") -migrations string The absolute path for the migrations folder (default "./migrations") -verify If set, no media will be imported and instead be tested to see if they've been imported already