Create Data Lake on AWS S3 to store dimensional tables after processing data using Spark on AWS EMR cluster
-
Updated
Oct 10, 2019 - Python
Create Data Lake on AWS S3 to store dimensional tables after processing data using Spark on AWS EMR cluster
Load data from the Million Song Dataset into a final dimensional model stored in S3.
Load data from the Million Song Dataset into AWS RedShift.
Add a description, image, and links to the dimensional-model topic page so that developers can more easily learn about it.
To associate your repository with the dimensional-model topic, visit your repo's landing page and select "manage topics."