Skip to content

[uber.com] Chaperone - система аудита Kafka, разработанная в Uber

License

Apache-2.0, Unknown licenses found

Licenses found

Apache-2.0
LICENSE
Unknown
LICENSE.txt
Notifications You must be signed in to change notification settings

gridgentoo/ChaperoneUber

Repository files navigation

Сборка из исходников Spark, Hive, Hadoop, Zookeper, Mesos, BigData, Tensorflow... в Gentoo

https://github.com/gridgentoo/gentoo-tensorflow-overlay

Как Uber масштабировал свою инфраструктуру в реальном времени до [триллиона событий] в день используя Экосистему Hadoop & Spark

Как Uber Engineering проводит аудит Kafka

[uber.com] Introducing Chaperone: How Uber Engineering Audits Kafka End-to-End [Chaperone] - это система аудита Kafka, разработанная в Uber

https://docs.google.com/document/d/1AvkvJjnbjGmMY0GnnwrCGPdhBEfEeG5khL1N9iYwQaE/

[github.com] Chaperone - это система аудита Kafka, разработанная в Uber

https://github.com/gridgentoo/ChaperoneUber

Image alt Image alt Image alt

[github.com] uReplicator: Uber Engineering’s Robust Kafka Replicator

https://github.com/gridgentoo/uReplicatorUber

[uber.com] uReplicator: Uber Engineering’s Robust Kafka Replicator https://docs.google.com/document/d/1jxMeu3ctc5UD2MfaurShfi1BiNwtLeUvzkw2VEye3II/edit#heading=h.v1nll5cd1ml6

Image alt Image alt Image alt

  • Как Uber масштабировал свою инфраструктуру в реальном времени до [триллиона событий] в день
  • How Uber scaled its Real Time Infrastructure to [Trillion events] per day

Chaperone

As Kafka audit system, Chaperone monitors the completeness and latency of data stream. The audit metrics are persisted in database for Kafka users to quantify the loss of their topics if any.

Basically, Chaperone cuts timeline into 10min buckets and assigns message to corresponding bucket according to its event time. The stats of the bucket are updated accordingly, like the total message count. Periodically, the stats are sent out to a dedicated Kafka topic, say 'chaperone-audit'. ChaperoneCollector consumes those stats from this topic and persists them into database.

============ Chaperone is made of several components:

  1. ChaperoneClient is a library that can be put in like Kafka Producer or Consumer to audit messages as they flow through. The audit stats are sent to a dedicated Kafka topic, say 'chaperone-audit'.
  2. ChaperoneCollector consumes audit stats from 'chaperone-audit' and persists them into database.
  3. ChaperoneService audits messages kept in Kafka. Since it's built upon uReplicator, it consists of two subsystems: ChaperoneServiceController to auto-detect topics in Kafka and assign the topic-partitions to workers to audit; ChaperoneServiceWorker to audit messages from assigned topic-partitions. In particular, ChaperoneService and ChaperoneCollector together ensure each message is audited exactly once.

Chaperone Quick Start

Get the Code

Check out the Chaperone project:

git clone git@github.com:uber/chaperone.git
cd chaperone

This project contains everything you’ll need to run Chaperone.

Build Chaperone

Before you can run Chaperone, you need to build a package for it.

mvn clean package

Or command below to skip tests

mvn clean package -DskipTests

Set Up Local Test Environment

To test Chaperone locally, you need two systems: Kafka, and ZooKeeper. The script “grid” is to help you set up these systems.

  • The command below will download, install, and start ZooKeeper and Kafka (named cluster1)
bin/grid bootstrap

Start ChaperoneService

  • Start ChaperoneService Controller
./ChaperoneDistribution/target/ChaperoneDistribution-pkg/bin/start-chaperone-controller.sh
  • Start ChaperoneService Worker
./ChaperoneDistribution/target/ChaperoneDistribution-pkg/bin/start-chaperone-worker.sh

Generate Load

  • Create a dummyTopic in Kafka and produce some dummy data:
./bin/produce-data-to-kafka-topic-dummyTopic.sh
  • Check if the data is successfully produced to Kafka by console-consumer as below:
./deploy/kafka/bin/kafka-console-consumer.sh --zookeeper localhost:2181/cluster1 --topic dummyTopic

You should get this data:

Kafka topic dummy topic data 1
Kafka topic dummy topic data 2
Kafka topic dummy topic data 3
Kafka topic dummy topic data 4
…

Check Audit Stats

In this example, the topic dummyTopic will be auto-detected and assigned to worker to audit. Periodically, the audit stats are sent to a topic called 'chaperone-audit'.

./deploy/kafka/bin/kafka-console-consumer.sh --zookeeper localhost:2181/cluster1 --topic chaperone-audit 

One can also manually add topic to audit by command below:

curl -X POST -d '{"topic":"dummyTopic", "numPartitions":"1"}' http://localhost:9000/topics

Start ChaperoneCollector

To start ChaperoneCollector, MySQL is required and Redis is optional. MySQL is used to persist audit stats and Redis is used to deduplicate. Deduplication can be turned off. The configuration file for ChaperoneCollector is ./config/chaperonecollector.properties, which might be updated to connect to MySQL and Redis.

  • Start ChaperoneCollector
./ChaperoneDistribution/target/ChaperoneDistribution-pkg/bin/start-chaperone-collector.sh

About

[uber.com] Chaperone - система аудита Kafka, разработанная в Uber

Resources

License

Apache-2.0, Unknown licenses found

Licenses found

Apache-2.0
LICENSE
Unknown
LICENSE.txt

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published