Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions docs/integrations/data-ingestion/clickpipes/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@
| Confluent Cloud | <Confluentsvg class="image" alt="Confluent Cloud logo" style={{width: '3rem'}}/> |Streaming| Stable | Unlock the combined power of Confluent and ClickHouse Cloud through our direct integration. |
| Redpanda | <Image img={redpanda_logo} size="logo" alt="Redpanda logo"/> |Streaming| Stable | Configure ClickPipes and start ingesting streaming data from Redpanda into ClickHouse Cloud. |
| AWS MSK | <Msksvg class="image" alt="AWS MSK logo" style={{width: '3rem', 'height': '3rem'}}/> |Streaming| Stable | Configure ClickPipes and start ingesting streaming data from AWS MSK into ClickHouse Cloud. |
| Azure Event Hubs | <Azureeventhubssvg class="image" alt="Azure Event Hubs logo" style={{width: '3rem'}}/> |Streaming| Stable | Configure ClickPipes and start ingesting streaming data from Azure Event Hubs into ClickHouse Cloud. |
| Azure Event Hubs | <Azureeventhubssvg class="image" alt="Azure Event Hubs logo" style={{width: '3rem'}}/> |Streaming| Stable | Configure ClickPipes and start ingesting streaming data from Azure Event Hubs into ClickHouse Cloud. See the [Azure Event Hubs FAQ](/integrations/clickpipes/kafka/faq/#azure-eventhubs) for guidance. |
| WarpStream | <Warpstreamsvg class="image" alt="WarpStream logo" style={{width: '3rem'}}/> |Streaming| Stable | Configure ClickPipes and start ingesting streaming data from WarpStream into ClickHouse Cloud. |
| Amazon S3 | <S3svg class="image" alt="Amazon S3 logo" style={{width: '3rem', height: 'auto'}}/> |Object Storage| Stable | Configure ClickPipes to ingest large volumes of data from object storage. |
| Google Cloud Storage | <Gcssvg class="image" alt="Google Cloud Storage logo" style={{width: '3rem', height: 'auto'}}/> |Object Storage| Stable | Configure ClickPipes to ingest large volumes of data from object storage. |
Expand Down Expand Up @@ -101,7 +101,7 @@

- **Does using ClickPipes incur an additional cost?**

ClickPipes is billed on two dimensions: Ingested Data and Compute. The full details of the pricing are available on [this page](/cloud/manage/jan-2025-faq/pricing-dimensions#clickpipes-pricing-faq). Running ClickPipes might also generate an indirect compute and storage cost on the destination ClickHouse Cloud service similar to any ingest workload.
ClickPipes is billed on two dimensions: Ingested Data and Compute. The full details of the pricing are available on [this page](/cloud/manage/billing/overview#clickpipes-for-streaming-object-storage). Running ClickPipes might also generate an indirect compute and storage cost on the destination ClickHouse Cloud service similar to any ingest workload.

Check warning on line 104 in docs/integrations/data-ingestion/clickpipes/index.md

View workflow job for this annotation

GitHub Actions / vale

ClickHouse.Colons

': I' should be in lowercase.

- **Is there a way to handle errors or failures when using ClickPipes for Kafka?**

Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
---
sidebar_label: 'Schema registries for Avro data streams'
description: 'Information on schema registries for Kafka ClickPipe'
sidebar_label: 'Integrate with a schema registry'
description: 'How to integrate for ClickPipes with a schema registry for schema management'
slug: /integrations/clickpipes/kafka/schema-registries
sidebar_position: 1
title: 'Schema registries for Kafka ClickPipe'
Expand All @@ -12,13 +12,12 @@ ClickPipes supports schema registries for Avro data streams.

## Supported registries for Kafka ClickPipes {#supported-schema-registries}

Schema registries that use the Confluent Schema Registry API are supported. This includes:
- Confluent Kafka and Cloud
- Redpanda
- AWS MSK
- Upstash
Schema registries that are API-compatible with the Confluent Schema Registry are supported. This includes:

ClickPipes is not currently compatible with the AWS Glue Schema registry or the Azure Schema Registry.
- Confluent Schema Registry
- Redpanda Schema Registry

ClickPipes does not support AWS Glue Schema Registry or Azure Schema Registry yet. If you require support for these schema registries, [reach out to our team](https://clickhouse.com/company/contact?loc=clickpipes).

## Configuration {#schema-registry-configuration}

Expand All @@ -31,7 +30,7 @@ ClickPipes with Avro data require a schema registry. This can be configured in o

## How it works {#how-schema-registries-work}

ClickPipes dynamically retrieves and applies the Avro schema from the configured Schema Registry.
ClickPipes dynamically retrieves and applies the Avro schema from the configured schema registry.
- If there's a schema id embedded in the message, it will use that to retrieve the schema.
- If there's no schema id embedded in the message, it will use the schema id or subject name specified in the ClickPipe configuration to retrieve the schema.
- If the message is written without an embedded schema id, and no schema id or subject name is specified in the ClickPipe configuration, then the schema will not be retrieved and the message will be skipped with a `SOURCE_SCHEMA_ERROR` logged in the ClickPipes errors table.
Expand Down
49 changes: 13 additions & 36 deletions docs/integrations/data-ingestion/clickpipes/kafka/05_faq.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
---
sidebar_label: 'FAQ'
description: 'Frequently asked questions about Kafka ClickPipes'
description: 'Frequently asked questions about ClickPipes for Kafka'
slug: /integrations/clickpipes/kafka/faq
sidebar_position: 1
title: 'Kafka ClickPipes FAQ'
Expand All @@ -23,7 +23,7 @@ ClickPipes uses a dedicated architecture running the Kafka Consumer API to read

The Kafka Table engine is a ClickHouse core capability that implements a "pull model" where the ClickHouse server itself connects to Kafka, pulls events then writes them locally.

ClickPipes is a separate cloud service that runs independently of the ClickHouse Service, it connects to Kafka (or other data sources) and pushes events to an associated ClickHouse Cloud service. This decoupled architecture allows for superior operational flexibility, clear separation of concerns, scalable ingestion, graceful failure management, extensibility and more.
ClickPipes is a separate cloud service that runs independently of the ClickHouse service. It connects to Kafka (or other data sources) and pushes events to an associated ClickHouse Cloud service. This decoupled architecture allows for superior operational flexibility, clear separation of concerns, scalable ingestion, graceful failure management, extensibility, and more.
</details>

<details>
Expand Down Expand Up @@ -63,77 +63,54 @@ Horizontal scaling adds more replicas to increase throughput, while vertical sca
This can be configured during ClickPipe creation, or at any other point under **Settings** -> **Advanced Settings** -> **Scaling**.
</details>

### Upstash {#upstash}

<details>

<summary>Does ClickPipes support Upstash?</summary>

Yes. The Upstash Kafka product entered into a deprecation period on 11th September 2024 for 6 months. Existing customers can continue to use ClickPipes with their existing Upstash Kafka brokers using the generic Kafka tile on the ClickPipes user interface. Existing Upstash Kafka ClickPipes are unaffected before the deprecation notice. When the the deprecation period is up the ClickPipe will stop functioning.
</details>

<details>

<summary>Does ClickPipes support Upstash schema registry?</summary>

No. ClickPipes is not Upstash Kafka schema registry compatible.
</details>

<details>

<summary>Does ClickPipes support the Upstash QStash Workflow?</summary>

No. Unless a Kafka compatible surface is introduced in QStash Workflow it will not work with Kafka ClickPipes.
</details>

### Azure EventHubs {#azure-eventhubs}
### Azure Event Hubs {#azure-eventhubs}

<details>

<summary>Does the Azure Event Hubs ClickPipe work without the Kafka surface?</summary>

No. ClickPipes requires the Azure Event Hubs to have the Kafka surface enabled. The Kafka protocol is supported for their Standard, Premium and Dedicated SKU only pricing tiers.
No. ClickPipes requires the Event Hubs namespace to have the Kafka surface enabled. This is only available in tiers above **basic**. See the [Azure Event Hubs documentation](https://learn.microsoft.com/en-us/azure/event-hubs/event-hubs-quickstart-kafka-enabled-event-hubs?tabs=passwordless#create-an-azure-event-hubs-namespace) for more information.
</details>

<details>

<summary>Does Azure schema registry work with ClickPipes</summary>
<summary>Does Azure Schema Registry work with ClickPipes?</summary>

No. ClickPipes is not currently Event Hubs Schema Registry compatible.
No. ClickPipes only supports schema registries that are API-compatible with the Confluent Schema Registry, which is not the case for Azure Schema Registry. If you require support for this schema registry, [reach out to our team](https://clickhouse.com/company/contact?loc=clickpipes).
</details>

<details>

<summary>What permissions does my policy need to consume from Azure Event Hubs?</summary>

To list topics and consume event, the shared access policy that is given to ClickPipes will at minimum require a 'Listen' claim.
To list topics and consume events, the shared access policy that is given to ClickPipes requires, at minimum, a 'Listen' claim.
</details>

<details>

<summary>Why is my Event Hubs not returning any data?</summary>

If your ClickHouse instance is in a different region or continent from your Event Hubs deployment, you may experience timeouts when onboarding your ClickPipes, and higher-latency when consuming data from the Event Hub. It is considered a best practice to locate your ClickHouse Cloud deployment and Azure Event Hubs deployment in cloud regions located close to each other to avoid adverse performance.
If your ClickHouse instance is in a different region or continent from your Event Hubs deployment, you may experience timeouts when onboarding your ClickPipes, and higher-latency when consuming data from the Event Hub. We recommend deploying ClickHouse Cloud and Azure Event Hubs in the same cloud region, or regions located close to each other, to avoid performance overhead.
</details>

<details>

<summary>Should I include the port number for Azure Event Hubs?</summary>

Yes. ClickPipes expects you to include your port number for the Kafka surface, which should be `:9093`.
Yes. ClickPipes expects you to include the port number for the Kafka surface, which should be `:9093`.
</details>

<details>

<summary>Are the ClickPipes IPs still relevant for Azure Event Hubs?</summary>
<summary>Are ClickPipes IPs still relevant for Azure Event Hubs?</summary>

Yes. If you restrict traffic to your Event Hubs instance please add the [documented static NAT IPs](../
/index.md#list-of-static-ips).
Yes. To restrict traffic to your Event Hubs instance, please add the [documented static NAT IPs](../
/index.md#list-of-static-ips) to .

</details>

<details>
<summary>Is the connection string for the Event Hub, or is it for the Event Hub namespace?</summary>

Both will work, however, we recommend using a shared access policy at the namespace level to retrieve samples from multiple Event Hubs.
Both work. We strongly recommend using a shared access policy at the **namespace level** to retrieve samples from multiple Event Hubs.
</details>
3 changes: 3 additions & 0 deletions scripts/aspell-ignore/en/aspell-dict.txt
Original file line number Diff line number Diff line change
Expand Up @@ -344,6 +344,7 @@ Doron
DoubleDelta
Doxygen
Draxlr
Dremio
Dresseler
Durre
ECMA
Expand Down Expand Up @@ -2638,6 +2639,7 @@ nats
navbar
ndjson
ness
Nessie
nestjs
netloc
newjson
Expand Down Expand Up @@ -3442,6 +3444,7 @@ trie
trimBoth
trimLeft
trimRight
Trino
trunc
tryBase
tryDecrypt
Expand Down
1 change: 1 addition & 0 deletions styles/ClickHouse/Headings.yml
Original file line number Diff line number Diff line change
Expand Up @@ -28,6 +28,7 @@ exceptions:
- Amazon Web Services
- Azure
- Azure Blob Storage
- Azure Event Hubs
- B2B
- BigQuery
- Bring Your Own Cloud
Expand Down