Skip to content
Switch branches/tags
Go to file
Latest commit 82e27d9 Jul 1, 2021 History
3 contributors

Users who have contributed to this file

@yihezkel @boazsha @ohadbitt

Quickstart: Ingestion from Kafka into Azure Data Explorer (Kusto) in Kafka Connect standalone mode

This is a quickstart for getting up and running with data ingestion from Apache Kafka into Azure Data Explorer (project code name Kusto) using the Kusto Sink Connector without having to deal with the complexities of Kafka cluster setup, creating a Kafka producer app, Kusto sink connector cluster setup.

The goal is to get started quickly, so all the requisite components for a Kafka ingestion pipeline into Kusto are self-contained and run in Docker containers - this includes a pseudo-distributed setup of Kafka, Zookeeper, Kafka Connect worker and the Kafka event generator/producer application.

Follow through the lab which leverages the storm events public dataset, and get a feel for the connector in isolation. You can then move to the more involved labs that cover distributed Kafka Connect, leveraging Azure PaaS and ISV IaaS Kafka offerings.

This lab is a contribution (thanks Abhishek Gupta - @abhirockzz) from the Cloud advocacy team - a team that strives to improve developer experience on Azure.

1. Prerequisites
2. Create an Azure Active Directory Service Principal
3. Provision and configure Azure Data Explorer
4. Clone the lab's git repo
5. Review contents
6. Start the lab
7. Check Azure Data Explorer for event delivery by the connector
8. Reset and Clean up

1. Prerequisites

2. Create an Azure Active Directory Service Principal

This service principal will be the identity used by the connector to write to the Azure Data Explorer table. In the next step, we will grant permissions for this service principal to access Azure Data Explorer.

2.1. Login to your Azure subscription via Azure CLI

az login

This launches a browser to authentciate. Follow the steps to authenticate.

2.2. Choose the subscription you want to run the lab in. This is needed when you have multiple.

az account set --subscription YOUR_SUBSCRIPTION_GUID

2.3. Create the service principal

Let's call our service principal, kusto-kafka-spn. Run the command below to create it.

az ad sp create-for-rbac -n "kusto-kafka-spn"

You will get a JSON response as shown below. Note the appId, password and tenant as you will need them in subsequent steps

  "appId": "fe7280c7-5705-4789-b17f-71a472340429",
  "displayName": "kusto-kafka-spn",
  "name": "http://kusto-kafka-spn",
  "password": "29c719dd-f2b3-46de-b71c-4004fb6116ee",
  "tenant": "42f988bf-86f1-42af-91ab-2d7cd011db42"

3. Provision and configure Azure Data Explorer

3.1. Create a cluster and database

3.2. Create a table and associated mapping

  1. Create a table called (Storms) and the corresponding table mapping to data needing ingesting (Storms_CSV_Mapping):
.create table Storms (StartTime: datetime, EndTime: datetime, EventId: int, State: string, EventType: string, Source: string)

.create table Storms ingestion csv mapping 'Storms_CSV_Mapping' '[{"Name":"StartTime","datatype":"datetime","Ordinal":0}, {"Name":"EndTime","datatype":"datetime","Ordinal":1},{"Name":"EventId","datatype":"int","Ordinal":2},{"Name":"State","datatype":"string","Ordinal":3},{"Name":"EventType","datatype":"string","Ordinal":4},{"Name":"Source","datatype":"string","Ordinal":5}]'

3.3. Create a batch ingestion policy on the table for configurable ingestion latency

The ingestion policy is a performance optimizer and includes three parameters, the first one met triggers an ingestion into Azure Data Explorer table.

.alter table Storms policy ingestionbatching @'{"MaximumBatchingTimeSpan":"00:00:15", "MaximumNumberOfItems": 100, "MaximumRawDataSizeMB": 300}'

3.4. Grant the service principal permission to work with the database

You will need the service principal details from section 2.3

.add database YOUR_DATABASE_NAME admins  ('aadapp=YOUR_APP_ID;YOUR_TENANT_ID') 'AAD App'

4. Clone the lab's git repo

  1. Create a local directory on your machine-
mkdir ~/kafka-kusto-hol
cd ~/kafka-kusto-hol
  1. Clone the repo-
cd ~/kafka-kusto-hol
git clone
cd azure-kusto-labs/kafka-integration/dockerized-quickstart

5. Review contents

5.1. List the contents

cd ~/kafka-kusto-hol/azure-kusto-labs/kafka-integration/dockerized-quickstart

This is what it should look like-

├── adx-query.png
├── adx-sink-config.json
├── connector
│   └── Dockerfile
├── docker-compose.yaml
└── storm-events-producer
    ├── Dockerfile
    ├── StormEvents.csv
    ├── go.mod
    ├── go.sum
    ├── kafka
    │   └── kafka.go
    └── main.go

5.2. adx-sink-config.json

This is the Kusto sink properties file we need to update with our specific configuration details for the lab.
Here is what it looks like-

    "name": "storm",
    "config": {
        "connector.class": "",
        "flush.size.bytes": 10000,
        "": 10000,
        "tasks.max": 1,
        "topics": "storm-events",
        "kusto.tables.topics.mapping": "[{'topic': 'storm-events','db': '<enter database name>', 'table': 'Storms','format': 'csv', 'mapping':'Storms_CSV_Mapping'}]",
        "aad.auth.authority": "<enter tenant ID>",
        "aad.auth.appid": "<enter application ID>",
        "aad.auth.appkey": "<enter client secret>",
        "kusto.ingestion.url": "https://ingest-<name of cluster>.<region>",
        "kusto.query.url": "https://<name of cluster>.<region>",
        "key.converter": "",
        "value.converter": ""

Replace the values for the following attributes per your Azure Data Explorer setup - aad.auth.authority, aad.auth.appid, aad.auth.appkey, kusto.tables.topics.mapping (the database name), kusto.ingestion.url and kusto.query.url.

5.3. connector/Dockerfile

Has the commands for generating the docker image for the connector instance. It includes download of the connector from the git repo release directory.

5.4. storm-events-producer directory and its contents

At a high level - this has a Go program that reads a local "StormEvents.csv" file and publishes it to a Kafka topic.

5.5. docker-compose.yaml

version: "2"
    image: debezium/zookeeper:1.2
      - 2181:2181
    image: debezium/kafka:1.2
      - 9092:9092
      - zookeeper
      - zookeeper
      - ZOOKEEPER_CONNECT=zookeeper:2181
      context: ./connector
      - 8083:8083
      - kafka
      - kafka
      - BOOTSTRAP_SERVERS=kafka:9092
      - GROUP_ID=adx
      - CONFIG_STORAGE_TOPIC=my_connect_configs
      - OFFSET_STORAGE_TOPIC=my_connect_offsets
      - STATUS_STORAGE_TOPIC=my_connect_statuses
      context: ./storm-events-producer
      - kafka
      - kafka
      - KAFKA_BOOTSTRAP_SERVER=kafka:9092
      - KAFKA_TOPIC=storm-events
      - SOURCE_FILE=StormEvents.csv

6: Start the lab

6.1. Start the containers - Kafka, connect, producer etc

In a terminal, start the containers-

docker-compose up

The producer application will start sending events to the storm-events topic. You should see logs similar to:

events-producer_1  | sent message to partition 0 offset 0
events-producer_1  | event  2007-01-01 00:00:00.0000000,2007-01-01 00:00:00.0000000,13208,NORTH CAROLINA,Thunderstorm Wind,Public
events-producer_1  | 
events-producer_1  | sent message to partition 0 offset 1
events-producer_1  | event  2007-01-01 00:00:00.0000000,2007-01-01 05:00:00.0000000,23358,WISCONSIN,Winter Storm,COOP Observer

Should you need to check the logs, in a separate terminal run the following-

docker-compose logs -f | grep kusto-connect

6.2. Start the connector via Kafka Connect REST call

In a separate terminal, launch sink task

curl -X POST -H "Content-Type: application/json" --data @adx-sink-config.json http://localhost:8083/connectors

Check status

curl http://localhost:8083/connectors/storm/status

The connector should start queueing ingestion processes to Azure Data Explorer.

7. Check Azure Data Explorer for event delivery by the connector

There may be a delay before data ends up in the Storms table. To confirm, check the row count and confirm that there are no failures in the ingestion process:

Storms | count

. show ingestion failures

Once there is some data, try out a few queries. To see all the records:


Use where and project to filter specific data

| where EventType == 'Drought' and State == 'TEXAS'
| project StartTime, EndTime, Source, EventId

Use the summarize operator

| summarize event_count=count() by State
| where event_count > 10
| project State, event_count
| render columnchart

These are just few examples. Dig into the Kusto Query Language documentation or explore tutorials about how to ingest JSON formatted sample data into Azure Data Explorer, using scalar operators, timecharts etc.

8. Reset and Clean up

If you want to re-start from scratch, simply stop the containers (docker-compose down -v), delete (drop table Storms) and re-create the Storms table (along with the mapping) and re-start containers (docker-compose up)

To delete the Azure Data Explorer cluster/database, use az cluster delete or az kusto database delete

az kusto cluster delete -n <cluster name> -g <resource group name>
az kusto database delete -n <database name> --cluster-name <cluster name> -g <resource group name>

If you want to only delete the connector

Curl - X DELETE http://localhost:8083/connectors/connector

This concludes the hands-on lab.