Skip to content

chronologos/roam-backup

 
 

Repository files navigation

[[roam-backup]]

Automate backing up JSON copies of your Roam Research data, using GitHub Actions and AWS S3/ Google Cloud Storage.

Backups will be generated using Puppeteer and saved into the backups folder (config.backupFolder, can be changed) of an S3 bucket of your choice.

Usage

You can get roam-backup up and running in just a few minutes! The best part is that you don't need to deploy anything!

Just follow these steps and you'll be on your way:

1. Fork this repository

If you haven't done this before, you'll find the Fork button on the top right of a GitHub repository's screen.

2. Enable Actions on your newly forked repository

This is necessary because Actions get disabled when you fork a repository. Do this by tapping on the "Actions" tab in your repository (next to "Pull Requests"), and hit the big green button.

3. Setup an AWS S3/ Google Cloud Storage bucket to store your Roam backups

AWS
  • Ensure you create the AWS S3 bucket manually –– this script will not create the bucket. Note the awsBucketName for step #4.
  • Create (or find an existing) user in the AWS console, and note its awsAccessKeyId and awsAccessKeySecret for step #4.
  • Ensure the user has permissions to upload to an S3 bucket. To do this, you'll need to attach a policy to the user that allows uploading to S3. The simplest way to do this would be to use the existing global policy: AmazonS3FullAccess.
  • (OPTIONAL) If you want to be a bit more conservative with the access policy, instead of AmazonS3FullAccess you can create your own custom policy and attach that to the user. Here's an example of the attributes you'd need to set:
    • Service: S3
    • Access Level: Write -> Put Object (NOTE: Do not simply select 'Write', instead click on the arrow to drill down into it and choose only 'Put Bucket' within all the options under Write)
    • Resource: Click on "Add ARN", and specify the Bucket Name, you can select "Any" for Object Name
    • Review and save the policy, then attach it to your user
  • Note the awsAccessKeyId and awsAccessKeySecret of the user for step #4.
GCP
  • If using Google Cloud Platform, set up a Google Cloud Storage Bucket as well a a service account authorized to use the bucket. Take note of the following: GCP_SA_KEY, GCP_SA_EMAIL, GCS_BUCKET_NAME.
  • I also like to set a retention policy to move to archival storage anything older than 14 days. This helps keep costs down.

4. Set your repository Secrets

Go to your Github repository's Settings tab, and click on Secrets section on the left. Add the following secrets (naming must match exactly!), using your Roam login credentials and the AWS bucket name and user access key ID/secret from step #3:

  • roamEmail
  • roamPassword
  • For AWS:
    • awsBucketName
    • awsAccessKeyId
    • awsAccessKeySecret
  • For GCP:
    • gcpSaKey
    • gcpSaEmail
    • gcsBucketName.
  • storageOpt should be set to either aws or gcp.

Don't worry! Your Roam, AWS and GCP credentials will be secure. GitHub Secrets are encrypted, and provide a convenient methodology for storing secure data in repositories.

5. Make a commit. It can be any commit, but this will start the process and trigger workflows.

Congrats! 🎉 You've successfully automated the backup of your brain 🧠. Now go write about it in today's Daily Note!

NOTE: This is still fairly WIP, and this is my first project using Puppeteer, so it may be a little buggy.

Development

Running this project locally should be possible using .env - copy .env.example to .env and fill it in with your own authentication keys.

The project generates an error.png screenshot to capture the current page if something goes wrong, as well as ZIP folders, which are the JSON backups. Running npm start will clear any local screenshots and backups, and run the script as it would in the GitHub Actions workflow (npm start)

About

Automated Roam Research backups using GitHub Actions and S3/GCS

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • JavaScript 100.0%