Raster Vision AWS Batch runner setup (for RV < 0.12)
This repository contains the deployment code that sets up the necessary AWS resources to utilize the AWS Batch runner in Raster Vision.
⚠️ For RV >= 0.12, the contents of this repo have moved into the main repo.
master branch of this repo should be used in conjunction with the
master branch (or
latest Docker image tag) of Raster Vision
which contains the latest changes. For versions of this repo that correspond to stable, released versions of Raster Vision, see:
Using Batch is advantageous because it starts and stops instances automatically and runs jobs sequentially or in parallel according to the dependencies between them. In addition, this deployment sets up distinct CPU and GPU resources and utilizes spot instances, which is more cost-effective than always using a GPU on-demand instance. Deployment is driven via the AWS console using a CloudFormation template. This AWS Batch setup is an "advanced" option that assumes some familiarity with Docker, AWS IAM, named profiles, availability zones, EC2, ECR, CloudFormation, and Batch.
Table of Contents
- AWS Account Setup
- AWS Credentials
- Deploying Batch resources
- Update Raster Vision configuration
- Deploy new job definitions
AWS Account Setup
In order to setup Batch using this repo, you will need to setup your AWS account so that:
- you have either root access to your AWS account, or an IAM user with admin permissions. It may be possible with less permissions, but we haven't figured out how to do this yet after some experimentation.
- you have the ability to launch P2 or P3 instances which have GPUs. In the past, it was necessary to open a support ticket to request access to these instances. You will know if this is the case if the Packer job fails when trying to launch the instance.
- you have requested permission from AWS to use availability zones outside the USA if you would like to use them. (New AWS accounts can't launch EC2 instances in other AZs by default.) If you are in doubt, just use us-east-1.
Using the AWS CLI, create an AWS profile for the target AWS environment. An example, naming the profile
$ aws --profile raster-vision configure AWS Access Key ID [****************F2DQ]: AWS Secret Access Key [****************TLJ/]: Default region name [us-east-1]: us-east-1 Default output format [None]:
You will be prompted to enter your AWS credentials, along with a default region. The Access Key ID and Secret Access Key can be retrieved from the IAM console. These credentials will be used to authenticate calls to the AWS API when using Packer and the AWS CLI.
Deploying Batch resources
To deploy AWS Batch resources using AWS CloudFormation, start by logging into your AWS console. Then, follow the steps below:
- Navigate to
CloudFormation > Create Stack
- In the
Choose a template field, select
Upload a template to Amazon S3and upload the template in
Prefix: If you are setting up multiple RV stacks within an AWS account, you need to set a prefix for namespacing resources. Otherwise, there will be name collisions with any resources that were created as part of another stack.
- Specify the following required parameters:
Stack Name: The name of your CloudFormation stack
VPC: The ID of the Virtual Private Cloud in which to deploy your resource. Your account should have at least one by default.
Subnets: The ID of any subnets that you want to deploy your resources into. Your account should have at least two by default; make sure that the subnets you select are in the VPC that you chose by using the AWS VPC console, or else CloudFormation will throw an error. (Subnets are tied to availability zones, and so affect spot prices.) In addition, you need to choose subnets that are available for the instance type you have chosen. To find which subnets are available, go to Spot Pricing History in the EC2 console and select the instance type. Then look up the availability zones that are present in the VPC console to find the corresponding subnets. Your spot requests will be more likely to be successful and your savings will be greater if you have subnets in more availability zones.
SSH Key Name: The name of the SSH key pair you want to be able to use to shell into your Batch instances. If you've created an EC2 instance before, you should already have one you can use; otherwise, you can create one in the EC2 console. Note: If you decide to create a new one, you will need to log out and then back in to the console before creating a Cloudformation stack using this key.
Instance Types: Provide the instance types you would like to use. (For GPUs,
p3.2xlargeis approximately 4 times the speed for 4 times the price.)
- Adjust any preset parameters that you want to change (the defaults should be fine for most users) and click
- Advanced users: If you plan on modifying Raster Vision and would like to publish a custom image to run on Batch, you will need to specify (Tensorflow CPU, Tensorflow GPU, and PyTorch) ECR repo names and a tag name to use for both. Note that the repo names cannot be the same as the Stack name (the first field in the UI) and cannot be the same as any existing ECR repo names. If you are in a team environment where you are sharing the AWS account, the repo names should contain an identifier such as your username.
- Accept all default options on the
Optionspage and click
I acknowledge that AWS CloudFormation might create IAM resources with custom nameson the
Reviewpage and click
- Watch your resources get deployed!
Optional: Publish local Raster Vision images to ECR
If you setup ECR repositories during the CloudFormation setup (the "advanced user" option), then you will need to follow this step, which publishes local Raster Vision images to those ECR repositories. Every time you make a change to your local Raster Vision images and want to use those on Batch, you will need to run this step.
./docker/build in the main Raster Vision repo to build local copies of the Tensorflow CPU, Tensorflow GPU, and PyTorch images.
settings.mk, fill out the options shown in the table below.
||The local Raster Vision TF CPU image to use.|
||The local Raster Vision TF GPU image to use.|
||The local Raster Vision PyTorch image to use.|
||The name of the ECR TF CPU image|
||The name of the ECR TF GPU image|
||The name of the ECR PYTORCH image|
||The ECR image tag to use, that is the tag in ECR_TF_CPU_IMAGE,ECR_TF_GPU_IMAGE, and ECR_PYTORCH_IMAGE|
make publish-images to publish the images to your ECR repositories.
Update Raster Vision configuration
Finally, make sure to update your Raster Vision configuration with the Batch resources that were created.
Deploy new job definitions
When a user starts working on a new RV-based project (or a new user starts working on an existing RV-based project), they will often want to publish a custom Docker image to ECR and use it when running on Batch. To facilitate this, there is a separate CloudFormation template for creating new job definitions. The idea is that for each user/project pair which is identified by a
Namespace string, a CPU and GPU job definition is created which point to a specified ECR repo using that
Namespace as the tag. After creating these new resources, the image should be published to
repo:namespace on ECR, and the new job definitions should be placed in a project-specific RV profile file.