Skip to content

Latest commit

 

History

History
53 lines (37 loc) · 1.23 KB

README.md

File metadata and controls

53 lines (37 loc) · 1.23 KB

Exploration of astroph dataset

Data

The data are collected by arxiv_collector

Deployment

We use AWS to host this project. For large files, we use the Amazon S3.

To upload the files, we first create a s3 bucket:

aws s3 mb s3://ml-bucket-3

Here ml-bucket-3 is the bucket name. We should also include this name in our config.json such that our script can get back our files on s3.

Virtual environment

pip install flask uuid zappa boto3 numpy tensorflow

The webapp is now 676 MB.

find . -name "tests" | xargs rm -rf
find . -name "*.pyc" -delete
find . -name "*.so" | xargs strip

We use zappa to deploy the package on the Amazon Lambda. We first run zappa init to initialize the project. Here are our sample configuration file:

{
    "dev": {
        "app_function": "webapp.app",
        "aws_region": "us-east-1",
        "profile_name": "default",
        "project_name": "webapp",
        "runtime": "python3.6",
        "s3_bucket": "YOUR-ZAPPA-BUCKET-NAME",
	    "slim_handler": true,
	    "use_precompiled_packages": false,
        "memory_size": 1024
    }
}

Note that we have added the last three options: slim_handler, use_precompiled_packages, and memory_size.