Skip to content

Latest commit

 

History

History

Dynamic Partitioning in Kinesis Data Firehose: Creating partitioning keys with inline parsing

firehose_dynamic_partition_with_inline

This is a kinesis data firehose dynamic partitioning project for Python development with CDK.

The cdk.json file tells the CDK Toolkit how to execute your app.

This project is set up like a standard Python project. The initialization process also creates a virtualenv within this project, stored under the .venv directory. To create the virtualenv it assumes that there is a python3 (or python for Windows) executable in your path with access to the venv package. If for any reason the automatic creation of the virtualenv fails, you can create the virtualenv manually.

To manually create a virtualenv on MacOS and Linux:

$ python3 -m venv .venv

After the init process completes and the virtualenv is created, you can use the following step to activate your virtualenv.

$ source .venv/bin/activate

If you are a Windows platform, you would activate the virtualenv like this:

% .venv\Scripts\activate.bat

Once the virtualenv is activated, you can install the required dependencies.

(.venv) $ pip install -r requirements.txt

At this point you can now synthesize the CloudFormation template for this code.

(.venv) $ export CDK_DEFAULT_ACCOUNT=$(aws sts get-caller-identity --query Account --output text)
(.venv) $ export CDK_DEFAULT_REGION=$(curl -s 169.254.169.254/latest/dynamic/instance-identity/document | jq -r .region)
(.venv) $ cdk synth \
  --parameters FirehoseStreamName='your-delivery-stream-name' \
  --parameters FirehoseDynamicPartitionKeys='your-jq-expressions-to-parse-dynamic-partition-keys' \
  --parameters FirehosePrefix='your-s3-bucket-prefix-for-dynamic-partitioning'

Creating partitioning keys with inline parsing

For example, let's define partitioning keys for it with inline parsing for the following sample data

{
   "type": {
    "device": "mobile",
    "event": "view"
  },
  "customer_id": "123456789012",
  "event_timestamp": 1565382027, #epoch timestamp
  "region": "us-east-1"
}

For example, you are going to choose to partition your data based on region, device, and event_timestamp. After selecting data parameters for your partitioning keys, you then map each parameter to a valid jq expression. The following table shows such a mapping of parameters to jq expressions:

Parameter jq expression
region .region
device .type.device
year .event_timestamp| strftime("%Y")
month .event_timestamp| strftime("%m")
day .event_timestamp| strftime("%d")
hour .event_timestamp| strftime("%H")

Now you can create kinesis data firehose with the above mapping of parameters to jq expressions like this:

(.venv) $ export CDK_DEFAULT_ACCOUNT=$(aws sts get-caller-identity --query Account --output text)
(.venv) $ export CDK_DEFAULT_REGION=$(curl -s 169.254.169.254/latest/dynamic/instance-identity/document | jq -r .region)
(.venv) $ cdk synth \
  --parameters FirehoseStreamName='PUT-S3-inline' \
  --parameters FirehoseDynamicPartitionKeys='{region:.region,device:.type.device,year:.event_timestamp|strftime("%Y"),month:.event_timestamp|strftime("%m"),day:.event_timestamp|strftime("%d"),hour:.event_timestamp|strftime("%H")}' \
  --parameters FirehosePrefix='region=!{partitionKeyFromQuery:region}/device=!{partitionKeyFromQuery:device}/year=!{partitionKeyFromQuery:year}/month=!{partitionKeyFromQuery:month}/day=!{partitionKeyFromQuery:day}/hour=!{partitionKeyFromQuery:hour}/'

Use cdk deploy command to create the stack shown above.

(.venv) $ cdk deploy \
  --parameters FirehoseStreamName='your-delivery-stream-name' \
  --parameters FirehoseDynamicPartitionKeys='your-jq-expressions-to-parse-dynamic-partition-keys' \
  --parameters FirehosePrefix='your-s3-bucket-prefix-for-dynamic-partitioning'

For example,

(.venv) $ cdk deploy \
  --parameters FirehoseStreamName='PUT-S3-inline' \
  --parameters FirehoseDynamicPartitionKeys='{region:.region,device:.type.device,year:.event_timestamp|strftime("%Y"),month:.event_timestamp|strftime("%m"),day:.event_timestamp|strftime("%d"),hour:.event_timestamp|strftime("%H")}' \
  --parameters FirehosePrefix='region=!{partitionKeyFromQuery:region}/device=!{partitionKeyFromQuery:device}/year=!{partitionKeyFromQuery:year}/month=!{partitionKeyFromQuery:month}/day=!{partitionKeyFromQuery:day}/hour=!{partitionKeyFromQuery:hour}/'

After cdk deploy has been completed, you can check dynamic partitioning configuration of kinesis data firehose as running the following command:

(.venv) $ aws firehose describe-delivery-stream \
  --delivery-stream-name 'your-delivery-stream-name' \
  | jq '.DeliveryStreamDescription.Destinations[0].ExtendedS3DestinationDescription'

For example,

(.venv) $ aws firehose describe-delivery-stream \
  --delivery-stream-name PUT-S3-inline \
  | jq '.DeliveryStreamDescription.Destinations[0].ExtendedS3DestinationDescription'

{
  "RoleARN": "arn:aws:iam::123456789012:role/KinesisFirehoseServiceRole-PUT-S3-inline-us-east-1",
  "BucketARN": "arn:aws:s3:::firehose-to-s3-us-east-1-we13viq",
  "Prefix": "region=!{partitionKeyFromQuery:region}/device=!{partitionKeyFromQuery:device}/year=!{partitionKeyFromQuery:year}/month=!{partitionKeyFromQuery:month}/day=!{partitionKeyFromQuery:day}/hour=!{partitionKeyFromQuery:hour}/",
  "ErrorOutputPrefix": "error/year=!{timestamp:yyyy}/month=!{timestamp:MM}/day=!{timestamp:dd}/hour=!{timestamp:HH}/!{firehose:error-output-type}",
  "BufferingHints": {
    "SizeInMBs": 128,
    "IntervalInSeconds": 60
  },
  "CompressionFormat": "UNCOMPRESSED",
  "EncryptionConfiguration": {
    "NoEncryptionConfig": "NoEncryption"
  },
  "CloudWatchLoggingOptions": {
    "Enabled": true,
    "LogGroupName": "/aws/kinesisfirehose/PUT-S3-inline",
    "LogStreamName": "DestinationDelivery"
  },
  "ProcessingConfiguration": {
    "Enabled": true,
    "Processors": [
      {
        "Type": "RecordDeAggregation",
        "Parameters": [
          {
            "ParameterName": "SubRecordType",
            "ParameterValue": "JSON"
          }
        ]
      },
      {
        "Type": "AppendDelimiterToRecord",
        "Parameters": []
      },
      {
        "Type": "MetadataExtraction",
        "Parameters": [
          {
            "ParameterName": "MetadataExtractionQuery",
            "ParameterValue": "{region:.region,device:.type.device,year:.event_timestamp|strftime(\"%Y\"),month:.event_timestamp|strftime(\"%m\"),day:.event_timestamp|strftime(\"%d\"),hour:.event_timestamp|strftime(\"%H\")}"
          },
          {
            "ParameterName": "JsonParsingEngine",
            "ParameterValue": "JQ-1.6"
          }
        ]
      }
    ]
  },
  "S3BackupMode": "Disabled",
  "DataFormatConversionConfiguration": {
    "Enabled": false
  }
}

Verify

If you want to generate sample data and verify it is being processed and stored as follows: Direct Put -> Kinesis Data Firehose -> S3,
Run gen_fake_firehose_data.py script on the EC2 instance by entering the following command:

(.venv) $ cd ..
(.venv) $ ls src/main/python/
gen_fake_firehose_data.py
(.venv) $ pip install boto3 Faker # pip install -r requirements.txt
(.venv) $ python src/main/python/gen_fake_firehose_data.py --stream-name 'your-delivery-stream-name' --max-count -1

If you would like to know more about the usage of this command, you can type

(.venv) $ python src/main/python/gen_fake_firehose_data.py --help

To add additional dependencies, for example other CDK libraries, just add them to your setup.py file and rerun the pip install -r requirements.txt command.

Clean Up

Delete the CloudFormation stack by running the below command.

(.venv) $ cdk destroy

Useful commands

  • cdk ls list all stacks in the app
  • cdk synth emits the synthesized CloudFormation template
  • cdk deploy deploy this stack to your default AWS account/region
  • cdk diff compare deployed stack with current state
  • cdk docs open CDK documentation

Learn more

Enjoy!