diff --git a/Jenkinsfile.s390x b/Jenkinsfile.s390x index 9d8f37b144998..819026eed0b3a 100644 --- a/Jenkinsfile.s390x +++ b/Jenkinsfile.s390x @@ -82,7 +82,7 @@ pipeline { steps { timeout(unit: 'HOURS', time: 7) { // Skip the test case execution of modules which are either not supported on s390x or vendor images are not available for s390x. - sh "./mvnw $MAVEN_PARAMS $MAVEN_TEST_PARAMS $MAVEN_TEST_LIMIT_PARAMS -Dmaven.test.failure.ignore=true -Dcheckstyle.skip=true verify -pl '!docs,!components/camel-kudu,!components/camel-djl,!components/camel-consul,!components/camel-pulsar,!components/camel-xmpp,!components/camel-google/camel-google-pubsub,!components/camel-hdfs,!components/camel-vertx/camel-vertx-kafka/camel-vertx-kafka-component,!components/camel-zookeeper,!components/camel-zookeeper-master'" + sh "./mvnw $MAVEN_PARAMS $MAVEN_TEST_PARAMS $MAVEN_TEST_LIMIT_PARAMS -Dmaven.test.failure.ignore=true -Dcheckstyle.skip=true verify -pl '!docs,!components/camel-kudu,!components/camel-djl,!components/camel-consul,!components/camel-pulsar,!components/camel-xmpp,!components/camel-google/camel-google-pubsub,!components/camel-hdfs,!components/camel-zookeeper,!components/camel-zookeeper-master'" } } post { diff --git a/bom/camel-bom/pom.xml b/bom/camel-bom/pom.xml index cfe98c11b6e87..f4e10404ea92a 100644 --- a/bom/camel-bom/pom.xml +++ b/bom/camel-bom/pom.xml @@ -46,11 +46,6 @@ camel-servicenow-maven-plugin ${project.version} - - org.apache.camel.maven - camel-vertx-kafka-maven-plugin - ${project.version} - org.apache.camel camel-activemq diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/StaticEndpointBuilders.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/StaticEndpointBuilders.java index bff8dd4b15c87..aae39789dbf63 100644 --- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/StaticEndpointBuilders.java +++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/StaticEndpointBuilders.java @@ -15227,55 +15227,6 @@ public static org.apache.camel.builder.endpoint.dsl.VertxHttpEndpointBuilderFact String path) { return org.apache.camel.builder.endpoint.dsl.VertxHttpEndpointBuilderFactory.endpointBuilder(componentName, path); } - /** - * Vert.x Kafka (camel-vertx-kafka) - * Sent and receive messages to/from an Apache Kafka broker using vert.x - * Kafka client - * - * Category: messaging - * Since: 3.7 - * Maven coordinates: org.apache.camel:camel-vertx-kafka - * - * Syntax: vertx-kafka:topic - * - * Path parameter: topic (required) - * Name of the topic to use. On the consumer you can use comma to separate - * multiple topics. A producer can only send a message to a single topic. - * - * @param path topic - * @return the dsl builder - */ - @Deprecated - public static org.apache.camel.builder.endpoint.dsl.VertxKafkaEndpointBuilderFactory.VertxKafkaEndpointBuilder vertxKafka( - String path) { - return org.apache.camel.builder.endpoint.dsl.VertxKafkaEndpointBuilderFactory.endpointBuilder("vertx-kafka", path); - } - /** - * Vert.x Kafka (camel-vertx-kafka) - * Sent and receive messages to/from an Apache Kafka broker using vert.x - * Kafka client - * - * Category: messaging - * Since: 3.7 - * Maven coordinates: org.apache.camel:camel-vertx-kafka - * - * Syntax: vertx-kafka:topic - * - * Path parameter: topic (required) - * Name of the topic to use. On the consumer you can use comma to separate - * multiple topics. A producer can only send a message to a single topic. - * - * @param componentName to use a custom component name for the endpoint - * instead of the default name - * @param path topic - * @return the dsl builder - */ - @Deprecated - public static org.apache.camel.builder.endpoint.dsl.VertxKafkaEndpointBuilderFactory.VertxKafkaEndpointBuilder vertxKafka( - String componentName, - String path) { - return org.apache.camel.builder.endpoint.dsl.VertxKafkaEndpointBuilderFactory.endpointBuilder(componentName, path); - } /** * Vert.x WebSocket (camel-vertx-websocket) * Expose WebSocket endpoints and connect to remote WebSocket servers using diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/VertxKafkaEndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/VertxKafkaEndpointBuilderFactory.java deleted file mode 100644 index ad42e1223f881..0000000000000 --- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/VertxKafkaEndpointBuilderFactory.java +++ /dev/null @@ -1,6318 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one or more - * contributor license agreements. See the NOTICE file distributed with - * this work for additional information regarding copyright ownership. - * The ASF licenses this file to You under the Apache License, Version 2.0 - * (the "License"); you may not use this file except in compliance with - * the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -package org.apache.camel.builder.endpoint.dsl; - -import java.util.*; -import java.util.Map; -import java.util.concurrent.*; -import java.util.function.*; -import java.util.stream.*; -import javax.annotation.Generated; -import org.apache.camel.builder.EndpointConsumerBuilder; -import org.apache.camel.builder.EndpointProducerBuilder; -import org.apache.camel.builder.endpoint.AbstractEndpointBuilder; - -/** - * Sent and receive messages to/from an Apache Kafka broker using vert.x Kafka - * client - * - * Generated by camel build tools - do NOT edit this file! - */ -@Generated("org.apache.camel.maven.packaging.EndpointDslMojo") -public interface VertxKafkaEndpointBuilderFactory { - - - /** - * Builder for endpoint consumers for the Vert.x Kafka component. - */ - public interface VertxKafkaEndpointConsumerBuilder - extends - EndpointConsumerBuilder { - default AdvancedVertxKafkaEndpointConsumerBuilder advanced() { - return (AdvancedVertxKafkaEndpointConsumerBuilder) this; - } - /** - * Sets additional properties for either kafka consumer or kafka - * producer in case they can't be set directly on the camel - * configurations (e.g: new Kafka properties that are not reflected yet - * in Camel configurations), the properties have to be prefixed with - * additionalProperties.. E.g: - * additionalProperties.transactional.id=12345&amp;additionalProperties.schema.registry.url=http://localhost:8811/avro. - * - * The option is a: <code>java.util.Map&lt;java.lang.String, - * java.lang.Object&gt;</code> type. - * The option is multivalued, and you can use the - * additionalProperties(String, Object) method to add a value (call the - * method multiple times to set more values). - * - * Group: common - * - * @param key the option key - * @param value the option value - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder additionalProperties( - String key, - Object value) { - doSetMultiValueProperty("additionalProperties", "additionalProperties." + key, value); - return this; - } - /** - * Sets additional properties for either kafka consumer or kafka - * producer in case they can't be set directly on the camel - * configurations (e.g: new Kafka properties that are not reflected yet - * in Camel configurations), the properties have to be prefixed with - * additionalProperties.. E.g: - * additionalProperties.transactional.id=12345&amp;additionalProperties.schema.registry.url=http://localhost:8811/avro. - * - * The option is a: <code>java.util.Map&lt;java.lang.String, - * java.lang.Object&gt;</code> type. - * The option is multivalued, and you can use the - * additionalProperties(String, Object) method to add a value (call the - * method multiple times to set more values). - * - * Group: common - * - * @param values the values - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder additionalProperties( - Map values) { - doSetMultiValueProperties("additionalProperties", "additionalProperties.", values); - return this; - } - /** - * A list of host/port pairs to use for establishing the initial - * connection to the Kafka cluster. The client will make use of all - * servers irrespective of which servers are specified here for - * bootstrapping&amp;mdash;this list only impacts the initial hosts - * used to discover the full set of servers. This list should be in the - * form host1:port1,host2:port2,.... Since these servers are just used - * for the initial connection to discover the full cluster membership - * (which may change dynamically), this list need not contain the full - * set of servers (you may want more than one, though, in case a server - * is down). - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param bootstrapServers the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder bootstrapServers( - String bootstrapServers) { - doSetProperty("bootstrapServers", bootstrapServers); - return this; - } - /** - * Controls how the client uses DNS lookups. If set to use_all_dns_ips, - * connect to each returned IP address in sequence until a successful - * connection is established. After a disconnection, the next IP is - * used. Once all IPs have been used once, the client resolves the IP(s) - * from the hostname again (both the JVM and the OS cache DNS name - * lookups, however). If set to - * resolve_canonical_bootstrap_servers_only, resolve each bootstrap - * address into a list of canonical names. After the bootstrap phase, - * this behaves the same as use_all_dns_ips. If set to default - * (deprecated), attempt to connect to the first IP address returned by - * the lookup, even if the lookup returns multiple IP addresses. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: use_all_dns_ips - * Group: common - * - * @param clientDnsLookup the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder clientDnsLookup( - String clientDnsLookup) { - doSetProperty("clientDnsLookup", clientDnsLookup); - return this; - } - /** - * An id string to pass to the server when making requests. The purpose - * of this is to be able to track the source of requests beyond just - * ip/port by allowing a logical application name to be included in - * server-side request logging. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param clientId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder clientId(String clientId) { - doSetProperty("clientId", clientId); - return this; - } - /** - * Close idle connections after the number of milliseconds specified by - * this config. - * - * The option is a: <code>long</code> type. - * - * Default: 9m - * Group: common - * - * @param connectionsMaxIdleMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder connectionsMaxIdleMs( - long connectionsMaxIdleMs) { - doSetProperty("connectionsMaxIdleMs", connectionsMaxIdleMs); - return this; - } - /** - * Close idle connections after the number of milliseconds specified by - * this config. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 9m - * Group: common - * - * @param connectionsMaxIdleMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder connectionsMaxIdleMs( - String connectionsMaxIdleMs) { - doSetProperty("connectionsMaxIdleMs", connectionsMaxIdleMs); - return this; - } - /** - * To use a custom HeaderFilterStrategy to filter header to and from - * Camel message. - * - * The option is a: - * <code>org.apache.camel.spi.HeaderFilterStrategy</code> - * type. - * - * Group: common - * - * @param headerFilterStrategy the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder headerFilterStrategy( - org.apache.camel.spi.HeaderFilterStrategy headerFilterStrategy) { - doSetProperty("headerFilterStrategy", headerFilterStrategy); - return this; - } - /** - * To use a custom HeaderFilterStrategy to filter header to and from - * Camel message. - * - * The option will be converted to a - * <code>org.apache.camel.spi.HeaderFilterStrategy</code> - * type. - * - * Group: common - * - * @param headerFilterStrategy the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder headerFilterStrategy( - String headerFilterStrategy) { - doSetProperty("headerFilterStrategy", headerFilterStrategy); - return this; - } - /** - * A list of classes to use as interceptors. Implementing the - * org.apache.kafka.clients.producer.ProducerInterceptor interface - * allows you to intercept (and possibly mutate) the records received by - * the producer before they are published to the Kafka cluster. By - * default, there are no interceptors. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param interceptorClasses the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder interceptorClasses( - String interceptorClasses) { - doSetProperty("interceptorClasses", interceptorClasses); - return this; - } - /** - * The period of time in milliseconds after which we force a refresh of - * metadata even if we haven't seen any partition leadership changes to - * proactively discover any new brokers or partitions. - * - * The option is a: <code>long</code> type. - * - * Default: 5m - * Group: common - * - * @param metadataMaxAgeMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder metadataMaxAgeMs( - long metadataMaxAgeMs) { - doSetProperty("metadataMaxAgeMs", metadataMaxAgeMs); - return this; - } - /** - * The period of time in milliseconds after which we force a refresh of - * metadata even if we haven't seen any partition leadership changes to - * proactively discover any new brokers or partitions. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 5m - * Group: common - * - * @param metadataMaxAgeMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder metadataMaxAgeMs( - String metadataMaxAgeMs) { - doSetProperty("metadataMaxAgeMs", metadataMaxAgeMs); - return this; - } - /** - * A list of classes to use as metrics reporters. Implementing the - * org.apache.kafka.common.metrics.MetricsReporter interface allows - * plugging in classes that will be notified of new metric creation. The - * JmxReporter is always included to register JMX statistics. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param metricReporters the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder metricReporters( - String metricReporters) { - doSetProperty("metricReporters", metricReporters); - return this; - } - /** - * The number of samples maintained to compute metrics. - * - * The option is a: <code>int</code> type. - * - * Default: 2 - * Group: common - * - * @param metricsNumSamples the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder metricsNumSamples( - int metricsNumSamples) { - doSetProperty("metricsNumSamples", metricsNumSamples); - return this; - } - /** - * The number of samples maintained to compute metrics. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 2 - * Group: common - * - * @param metricsNumSamples the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder metricsNumSamples( - String metricsNumSamples) { - doSetProperty("metricsNumSamples", metricsNumSamples); - return this; - } - /** - * The highest recording level for metrics. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: INFO - * Group: common - * - * @param metricsRecordingLevel the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder metricsRecordingLevel( - String metricsRecordingLevel) { - doSetProperty("metricsRecordingLevel", metricsRecordingLevel); - return this; - } - /** - * The window of time a metrics sample is computed over. - * - * The option is a: <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param metricsSampleWindowMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder metricsSampleWindowMs( - long metricsSampleWindowMs) { - doSetProperty("metricsSampleWindowMs", metricsSampleWindowMs); - return this; - } - /** - * The window of time a metrics sample is computed over. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param metricsSampleWindowMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder metricsSampleWindowMs( - String metricsSampleWindowMs) { - doSetProperty("metricsSampleWindowMs", metricsSampleWindowMs); - return this; - } - /** - * The partition to which the record will be sent (or null if no - * partition was specified) or read from a particular partition if set. - * Header VertxKafkaConstants#PARTITION_ID If configured, it will take - * precedence over this config. - * - * The option is a: <code>java.lang.Integer</code> type. - * - * Group: common - * - * @param partitionId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder partitionId( - Integer partitionId) { - doSetProperty("partitionId", partitionId); - return this; - } - /** - * The partition to which the record will be sent (or null if no - * partition was specified) or read from a particular partition if set. - * Header VertxKafkaConstants#PARTITION_ID If configured, it will take - * precedence over this config. - * - * The option will be converted to a - * <code>java.lang.Integer</code> type. - * - * Group: common - * - * @param partitionId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder partitionId(String partitionId) { - doSetProperty("partitionId", partitionId); - return this; - } - /** - * The size of the TCP receive buffer (SO_RCVBUF) to use when reading - * data. If the value is -1, the OS default will be used. - * - * The option is a: <code>int</code> type. - * - * Default: 32768 - * Group: common - * - * @param receiveBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder receiveBufferBytes( - int receiveBufferBytes) { - doSetProperty("receiveBufferBytes", receiveBufferBytes); - return this; - } - /** - * The size of the TCP receive buffer (SO_RCVBUF) to use when reading - * data. If the value is -1, the OS default will be used. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 32768 - * Group: common - * - * @param receiveBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder receiveBufferBytes( - String receiveBufferBytes) { - doSetProperty("receiveBufferBytes", receiveBufferBytes); - return this; - } - /** - * The maximum amount of time in milliseconds to wait when reconnecting - * to a broker that has repeatedly failed to connect. If provided, the - * backoff per host will increase exponentially for each consecutive - * connection failure, up to this maximum. After calculating the backoff - * increase, 20% random jitter is added to avoid connection storms. - * - * The option is a: <code>long</code> type. - * - * Default: 1s - * Group: common - * - * @param reconnectBackoffMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder reconnectBackoffMaxMs( - long reconnectBackoffMaxMs) { - doSetProperty("reconnectBackoffMaxMs", reconnectBackoffMaxMs); - return this; - } - /** - * The maximum amount of time in milliseconds to wait when reconnecting - * to a broker that has repeatedly failed to connect. If provided, the - * backoff per host will increase exponentially for each consecutive - * connection failure, up to this maximum. After calculating the backoff - * increase, 20% random jitter is added to avoid connection storms. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 1s - * Group: common - * - * @param reconnectBackoffMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder reconnectBackoffMaxMs( - String reconnectBackoffMaxMs) { - doSetProperty("reconnectBackoffMaxMs", reconnectBackoffMaxMs); - return this; - } - /** - * The base amount of time to wait before attempting to reconnect to a - * given host. This avoids repeatedly connecting to a host in a tight - * loop. This backoff applies to all connection attempts by the client - * to a broker. - * - * The option is a: <code>long</code> type. - * - * Default: 50ms - * Group: common - * - * @param reconnectBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder reconnectBackoffMs( - long reconnectBackoffMs) { - doSetProperty("reconnectBackoffMs", reconnectBackoffMs); - return this; - } - /** - * The base amount of time to wait before attempting to reconnect to a - * given host. This avoids repeatedly connecting to a host in a tight - * loop. This backoff applies to all connection attempts by the client - * to a broker. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 50ms - * Group: common - * - * @param reconnectBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder reconnectBackoffMs( - String reconnectBackoffMs) { - doSetProperty("reconnectBackoffMs", reconnectBackoffMs); - return this; - } - /** - * The configuration controls the maximum amount of time the client will - * wait for the response of a request. If the response is not received - * before the timeout elapses the client will resend the request if - * necessary or fail the request if retries are exhausted. This should - * be larger than replica.lag.time.max.ms (a broker configuration) to - * reduce the possibility of message duplication due to unnecessary - * producer retries. - * - * The option is a: <code>int</code> type. - * - * Default: 30s - * Group: common - * - * @param requestTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder requestTimeoutMs( - int requestTimeoutMs) { - doSetProperty("requestTimeoutMs", requestTimeoutMs); - return this; - } - /** - * The configuration controls the maximum amount of time the client will - * wait for the response of a request. If the response is not received - * before the timeout elapses the client will resend the request if - * necessary or fail the request if retries are exhausted. This should - * be larger than replica.lag.time.max.ms (a broker configuration) to - * reduce the possibility of message duplication due to unnecessary - * producer retries. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 30s - * Group: common - * - * @param requestTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder requestTimeoutMs( - String requestTimeoutMs) { - doSetProperty("requestTimeoutMs", requestTimeoutMs); - return this; - } - /** - * The amount of time to wait before attempting to retry a failed - * request to a given topic partition. This avoids repeatedly sending - * requests in a tight loop under some failure scenarios. - * - * The option is a: <code>long</code> type. - * - * Default: 100ms - * Group: common - * - * @param retryBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder retryBackoffMs( - long retryBackoffMs) { - doSetProperty("retryBackoffMs", retryBackoffMs); - return this; - } - /** - * The amount of time to wait before attempting to retry a failed - * request to a given topic partition. This avoids repeatedly sending - * requests in a tight loop under some failure scenarios. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 100ms - * Group: common - * - * @param retryBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder retryBackoffMs( - String retryBackoffMs) { - doSetProperty("retryBackoffMs", retryBackoffMs); - return this; - } - /** - * The size of the TCP send buffer (SO_SNDBUF) to use when sending data. - * If the value is -1, the OS default will be used. - * - * The option is a: <code>int</code> type. - * - * Default: 131072 - * Group: common - * - * @param sendBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sendBufferBytes( - int sendBufferBytes) { - doSetProperty("sendBufferBytes", sendBufferBytes); - return this; - } - /** - * The size of the TCP send buffer (SO_SNDBUF) to use when sending data. - * If the value is -1, the OS default will be used. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 131072 - * Group: common - * - * @param sendBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sendBufferBytes( - String sendBufferBytes) { - doSetProperty("sendBufferBytes", sendBufferBytes); - return this; - } - /** - * The maximum amount of time the client will wait for the socket - * connection to be established. The connection setup timeout will - * increase exponentially for each consecutive connection failure up to - * this maximum. To avoid connection storms, a randomization factor of - * 0.2 will be applied to the timeout resulting in a random range - * between 20% below and 20% above the computed value. - * - * The option is a: <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param socketConnectionSetupTimeoutMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder socketConnectionSetupTimeoutMaxMs( - long socketConnectionSetupTimeoutMaxMs) { - doSetProperty("socketConnectionSetupTimeoutMaxMs", socketConnectionSetupTimeoutMaxMs); - return this; - } - /** - * The maximum amount of time the client will wait for the socket - * connection to be established. The connection setup timeout will - * increase exponentially for each consecutive connection failure up to - * this maximum. To avoid connection storms, a randomization factor of - * 0.2 will be applied to the timeout resulting in a random range - * between 20% below and 20% above the computed value. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param socketConnectionSetupTimeoutMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder socketConnectionSetupTimeoutMaxMs( - String socketConnectionSetupTimeoutMaxMs) { - doSetProperty("socketConnectionSetupTimeoutMaxMs", socketConnectionSetupTimeoutMaxMs); - return this; - } - /** - * The amount of time the client will wait for the socket connection to - * be established. If the connection is not built before the timeout - * elapses, clients will close the socket channel. - * - * The option is a: <code>long</code> type. - * - * Default: 10s - * Group: common - * - * @param socketConnectionSetupTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder socketConnectionSetupTimeoutMs( - long socketConnectionSetupTimeoutMs) { - doSetProperty("socketConnectionSetupTimeoutMs", socketConnectionSetupTimeoutMs); - return this; - } - /** - * The amount of time the client will wait for the socket connection to - * be established. If the connection is not built before the timeout - * elapses, clients will close the socket channel. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 10s - * Group: common - * - * @param socketConnectionSetupTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder socketConnectionSetupTimeoutMs( - String socketConnectionSetupTimeoutMs) { - doSetProperty("socketConnectionSetupTimeoutMs", socketConnectionSetupTimeoutMs); - return this; - } - /** - * Allow automatic topic creation on the broker when subscribing to or - * assigning a topic. A topic being subscribed to will be automatically - * created only if the broker allows for it using - * auto.create.topics.enable broker configuration. This configuration - * must be set to false when using brokers older than 0.11.0. - * - * The option is a: <code>boolean</code> type. - * - * Default: true - * Group: consumer - * - * @param allowAutoCreateTopics the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder allowAutoCreateTopics( - boolean allowAutoCreateTopics) { - doSetProperty("allowAutoCreateTopics", allowAutoCreateTopics); - return this; - } - /** - * Allow automatic topic creation on the broker when subscribing to or - * assigning a topic. A topic being subscribed to will be automatically - * created only if the broker allows for it using - * auto.create.topics.enable broker configuration. This configuration - * must be set to false when using brokers older than 0.11.0. - * - * The option will be converted to a <code>boolean</code> - * type. - * - * Default: true - * Group: consumer - * - * @param allowAutoCreateTopics the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder allowAutoCreateTopics( - String allowAutoCreateTopics) { - doSetProperty("allowAutoCreateTopics", allowAutoCreateTopics); - return this; - } - /** - * Whether to allow doing manual commits via - * org.apache.camel.component.vertx.kafka.offset.VertxKafkaManualCommit. - * If this option is enabled then an instance of - * org.apache.camel.component.vertx.kafka.offset.VertxKafkaManualCommit - * is stored on the Exchange message header, which allows end users to - * access this API and perform manual offset commits via the Kafka - * consumer. Note: To take full control of the offset committing, you - * may need to disable the Kafka Consumer default auto commit behavior - * by setting 'enableAutoCommit' to 'false'. - * - * The option is a: <code>boolean</code> type. - * - * Default: false - * Group: consumer - * - * @param allowManualCommit the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder allowManualCommit( - boolean allowManualCommit) { - doSetProperty("allowManualCommit", allowManualCommit); - return this; - } - /** - * Whether to allow doing manual commits via - * org.apache.camel.component.vertx.kafka.offset.VertxKafkaManualCommit. - * If this option is enabled then an instance of - * org.apache.camel.component.vertx.kafka.offset.VertxKafkaManualCommit - * is stored on the Exchange message header, which allows end users to - * access this API and perform manual offset commits via the Kafka - * consumer. Note: To take full control of the offset committing, you - * may need to disable the Kafka Consumer default auto commit behavior - * by setting 'enableAutoCommit' to 'false'. - * - * The option will be converted to a <code>boolean</code> - * type. - * - * Default: false - * Group: consumer - * - * @param allowManualCommit the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder allowManualCommit( - String allowManualCommit) { - doSetProperty("allowManualCommit", allowManualCommit); - return this; - } - /** - * The frequency in milliseconds that the consumer offsets are - * auto-committed to Kafka if enable.auto.commit is set to true. - * - * The option is a: <code>int</code> type. - * - * Default: 5s - * Group: consumer - * - * @param autoCommitIntervalMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder autoCommitIntervalMs( - int autoCommitIntervalMs) { - doSetProperty("autoCommitIntervalMs", autoCommitIntervalMs); - return this; - } - /** - * The frequency in milliseconds that the consumer offsets are - * auto-committed to Kafka if enable.auto.commit is set to true. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 5s - * Group: consumer - * - * @param autoCommitIntervalMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder autoCommitIntervalMs( - String autoCommitIntervalMs) { - doSetProperty("autoCommitIntervalMs", autoCommitIntervalMs); - return this; - } - /** - * What to do when there is no initial offset in Kafka or if the current - * offset does not exist any more on the server (e.g. because that data - * has been deleted): earliest: automatically reset the offset to the - * earliest offsetlatest: automatically reset the offset to the latest - * offsetnone: throw exception to the consumer if no previous offset is - * found for the consumer's groupanything else: throw exception to the - * consumer. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: latest - * Group: consumer - * - * @param autoOffsetReset the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder autoOffsetReset( - String autoOffsetReset) { - doSetProperty("autoOffsetReset", autoOffsetReset); - return this; - } - /** - * Automatically check the CRC32 of the records consumed. This ensures - * no on-the-wire or on-disk corruption to the messages occurred. This - * check adds some overhead, so it may be disabled in cases seeking - * extreme performance. - * - * The option is a: <code>boolean</code> type. - * - * Default: true - * Group: consumer - * - * @param checkCrcs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder checkCrcs(boolean checkCrcs) { - doSetProperty("checkCrcs", checkCrcs); - return this; - } - /** - * Automatically check the CRC32 of the records consumed. This ensures - * no on-the-wire or on-disk corruption to the messages occurred. This - * check adds some overhead, so it may be disabled in cases seeking - * extreme performance. - * - * The option will be converted to a <code>boolean</code> - * type. - * - * Default: true - * Group: consumer - * - * @param checkCrcs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder checkCrcs(String checkCrcs) { - doSetProperty("checkCrcs", checkCrcs); - return this; - } - /** - * A rack identifier for this client. This can be any string value which - * indicates where this client is physically located. It corresponds - * with the broker config 'broker.rack'. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: consumer - * - * @param clientRack the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder clientRack(String clientRack) { - doSetProperty("clientRack", clientRack); - return this; - } - /** - * Specifies the timeout (in milliseconds) for client APIs. This - * configuration is used as the default timeout for all client - * operations that do not specify a timeout parameter. - * - * The option is a: <code>int</code> type. - * - * Default: 1m - * Group: consumer - * - * @param defaultApiTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder defaultApiTimeoutMs( - int defaultApiTimeoutMs) { - doSetProperty("defaultApiTimeoutMs", defaultApiTimeoutMs); - return this; - } - /** - * Specifies the timeout (in milliseconds) for client APIs. This - * configuration is used as the default timeout for all client - * operations that do not specify a timeout parameter. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 1m - * Group: consumer - * - * @param defaultApiTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder defaultApiTimeoutMs( - String defaultApiTimeoutMs) { - doSetProperty("defaultApiTimeoutMs", defaultApiTimeoutMs); - return this; - } - /** - * If true the consumer's offset will be periodically committed in the - * background. - * - * The option is a: <code>boolean</code> type. - * - * Default: true - * Group: consumer - * - * @param enableAutoCommit the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder enableAutoCommit( - boolean enableAutoCommit) { - doSetProperty("enableAutoCommit", enableAutoCommit); - return this; - } - /** - * If true the consumer's offset will be periodically committed in the - * background. - * - * The option will be converted to a <code>boolean</code> - * type. - * - * Default: true - * Group: consumer - * - * @param enableAutoCommit the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder enableAutoCommit( - String enableAutoCommit) { - doSetProperty("enableAutoCommit", enableAutoCommit); - return this; - } - /** - * Whether internal topics matching a subscribed pattern should be - * excluded from the subscription. It is always possible to explicitly - * subscribe to an internal topic. - * - * The option is a: <code>boolean</code> type. - * - * Default: true - * Group: consumer - * - * @param excludeInternalTopics the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder excludeInternalTopics( - boolean excludeInternalTopics) { - doSetProperty("excludeInternalTopics", excludeInternalTopics); - return this; - } - /** - * Whether internal topics matching a subscribed pattern should be - * excluded from the subscription. It is always possible to explicitly - * subscribe to an internal topic. - * - * The option will be converted to a <code>boolean</code> - * type. - * - * Default: true - * Group: consumer - * - * @param excludeInternalTopics the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder excludeInternalTopics( - String excludeInternalTopics) { - doSetProperty("excludeInternalTopics", excludeInternalTopics); - return this; - } - /** - * The maximum amount of data the server should return for a fetch - * request. Records are fetched in batches by the consumer, and if the - * first record batch in the first non-empty partition of the fetch is - * larger than this value, the record batch will still be returned to - * ensure that the consumer can make progress. As such, this is not a - * absolute maximum. The maximum record batch size accepted by the - * broker is defined via message.max.bytes (broker config) or - * max.message.bytes (topic config). Note that the consumer performs - * multiple fetches in parallel. - * - * The option is a: <code>int</code> type. - * - * Default: 52428800 - * Group: consumer - * - * @param fetchMaxBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder fetchMaxBytes( - int fetchMaxBytes) { - doSetProperty("fetchMaxBytes", fetchMaxBytes); - return this; - } - /** - * The maximum amount of data the server should return for a fetch - * request. Records are fetched in batches by the consumer, and if the - * first record batch in the first non-empty partition of the fetch is - * larger than this value, the record batch will still be returned to - * ensure that the consumer can make progress. As such, this is not a - * absolute maximum. The maximum record batch size accepted by the - * broker is defined via message.max.bytes (broker config) or - * max.message.bytes (topic config). Note that the consumer performs - * multiple fetches in parallel. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 52428800 - * Group: consumer - * - * @param fetchMaxBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder fetchMaxBytes( - String fetchMaxBytes) { - doSetProperty("fetchMaxBytes", fetchMaxBytes); - return this; - } - /** - * The maximum amount of time the server will block before answering the - * fetch request if there isn't sufficient data to immediately satisfy - * the requirement given by fetch.min.bytes. - * - * The option is a: <code>int</code> type. - * - * Default: 500ms - * Group: consumer - * - * @param fetchMaxWaitMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder fetchMaxWaitMs( - int fetchMaxWaitMs) { - doSetProperty("fetchMaxWaitMs", fetchMaxWaitMs); - return this; - } - /** - * The maximum amount of time the server will block before answering the - * fetch request if there isn't sufficient data to immediately satisfy - * the requirement given by fetch.min.bytes. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 500ms - * Group: consumer - * - * @param fetchMaxWaitMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder fetchMaxWaitMs( - String fetchMaxWaitMs) { - doSetProperty("fetchMaxWaitMs", fetchMaxWaitMs); - return this; - } - /** - * The minimum amount of data the server should return for a fetch - * request. If insufficient data is available the request will wait for - * that much data to accumulate before answering the request. The - * default setting of 1 byte means that fetch requests are answered as - * soon as a single byte of data is available or the fetch request times - * out waiting for data to arrive. Setting this to something greater - * than 1 will cause the server to wait for larger amounts of data to - * accumulate which can improve server throughput a bit at the cost of - * some additional latency. - * - * The option is a: <code>int</code> type. - * - * Default: 1 - * Group: consumer - * - * @param fetchMinBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder fetchMinBytes( - int fetchMinBytes) { - doSetProperty("fetchMinBytes", fetchMinBytes); - return this; - } - /** - * The minimum amount of data the server should return for a fetch - * request. If insufficient data is available the request will wait for - * that much data to accumulate before answering the request. The - * default setting of 1 byte means that fetch requests are answered as - * soon as a single byte of data is available or the fetch request times - * out waiting for data to arrive. Setting this to something greater - * than 1 will cause the server to wait for larger amounts of data to - * accumulate which can improve server throughput a bit at the cost of - * some additional latency. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 1 - * Group: consumer - * - * @param fetchMinBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder fetchMinBytes( - String fetchMinBytes) { - doSetProperty("fetchMinBytes", fetchMinBytes); - return this; - } - /** - * A unique string that identifies the consumer group this consumer - * belongs to. This property is required if the consumer uses either the - * group management functionality by using subscribe(topic) or the - * Kafka-based offset management strategy. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: consumer - * - * @param groupId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder groupId(String groupId) { - doSetProperty("groupId", groupId); - return this; - } - /** - * A unique identifier of the consumer instance provided by the end - * user. Only non-empty strings are permitted. If set, the consumer is - * treated as a static member, which means that only one instance with - * this ID is allowed in the consumer group at any time. This can be - * used in combination with a larger session timeout to avoid group - * rebalances caused by transient unavailability (e.g. process - * restarts). If not set, the consumer will join the group as a dynamic - * member, which is the traditional behavior. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: consumer - * - * @param groupInstanceId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder groupInstanceId( - String groupInstanceId) { - doSetProperty("groupInstanceId", groupInstanceId); - return this; - } - /** - * The expected time between heartbeats to the consumer coordinator when - * using Kafka's group management facilities. Heartbeats are used to - * ensure that the consumer's session stays active and to facilitate - * rebalancing when new consumers join or leave the group. The value - * must be set lower than session.timeout.ms, but typically should be - * set no higher than 1/3 of that value. It can be adjusted even lower - * to control the expected time for normal rebalances. - * - * The option is a: <code>int</code> type. - * - * Default: 3s - * Group: consumer - * - * @param heartbeatIntervalMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder heartbeatIntervalMs( - int heartbeatIntervalMs) { - doSetProperty("heartbeatIntervalMs", heartbeatIntervalMs); - return this; - } - /** - * The expected time between heartbeats to the consumer coordinator when - * using Kafka's group management facilities. Heartbeats are used to - * ensure that the consumer's session stays active and to facilitate - * rebalancing when new consumers join or leave the group. The value - * must be set lower than session.timeout.ms, but typically should be - * set no higher than 1/3 of that value. It can be adjusted even lower - * to control the expected time for normal rebalances. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 3s - * Group: consumer - * - * @param heartbeatIntervalMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder heartbeatIntervalMs( - String heartbeatIntervalMs) { - doSetProperty("heartbeatIntervalMs", heartbeatIntervalMs); - return this; - } - /** - * Controls how to read messages written transactionally. If set to - * read_committed, consumer.poll() will only return transactional - * messages which have been committed. If set to read_uncommitted (the - * default), consumer.poll() will return all messages, even - * transactional messages which have been aborted. Non-transactional - * messages will be returned unconditionally in either mode. Messages - * will always be returned in offset order. Hence, in read_committed - * mode, consumer.poll() will only return messages up to the last stable - * offset (LSO), which is the one less than the offset of the first open - * transaction. In particular any messages appearing after messages - * belonging to ongoing transactions will be withheld until the relevant - * transaction has been completed. As a result, read_committed consumers - * will not be able to read up to the high watermark when there are in - * flight transactions. Further, when in read_committed the seekToEnd - * method will return the LSO. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: read_uncommitted - * Group: consumer - * - * @param isolationLevel the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder isolationLevel( - String isolationLevel) { - doSetProperty("isolationLevel", isolationLevel); - return this; - } - /** - * Deserializer class for key that implements the - * org.apache.kafka.common.serialization.Deserializer interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: org.apache.kafka.common.serialization.StringDeserializer - * Group: consumer - * - * @param keyDeserializer the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder keyDeserializer( - String keyDeserializer) { - doSetProperty("keyDeserializer", keyDeserializer); - return this; - } - /** - * The maximum amount of data per-partition the server will return. - * Records are fetched in batches by the consumer. If the first record - * batch in the first non-empty partition of the fetch is larger than - * this limit, the batch will still be returned to ensure that the - * consumer can make progress. The maximum record batch size accepted by - * the broker is defined via message.max.bytes (broker config) or - * max.message.bytes (topic config). See fetch.max.bytes for limiting - * the consumer request size. - * - * The option is a: <code>int</code> type. - * - * Default: 1048576 - * Group: consumer - * - * @param maxPartitionFetchBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder maxPartitionFetchBytes( - int maxPartitionFetchBytes) { - doSetProperty("maxPartitionFetchBytes", maxPartitionFetchBytes); - return this; - } - /** - * The maximum amount of data per-partition the server will return. - * Records are fetched in batches by the consumer. If the first record - * batch in the first non-empty partition of the fetch is larger than - * this limit, the batch will still be returned to ensure that the - * consumer can make progress. The maximum record batch size accepted by - * the broker is defined via message.max.bytes (broker config) or - * max.message.bytes (topic config). See fetch.max.bytes for limiting - * the consumer request size. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 1048576 - * Group: consumer - * - * @param maxPartitionFetchBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder maxPartitionFetchBytes( - String maxPartitionFetchBytes) { - doSetProperty("maxPartitionFetchBytes", maxPartitionFetchBytes); - return this; - } - /** - * The maximum delay between invocations of poll() when using consumer - * group management. This places an upper bound on the amount of time - * that the consumer can be idle before fetching more records. If poll() - * is not called before expiration of this timeout, then the consumer is - * considered failed and the group will rebalance in order to reassign - * the partitions to another member. For consumers using a non-null - * group.instance.id which reach this timeout, partitions will not be - * immediately reassigned. Instead, the consumer will stop sending - * heartbeats and partitions will be reassigned after expiration of - * session.timeout.ms. This mirrors the behavior of a static consumer - * which has shutdown. - * - * The option is a: <code>int</code> type. - * - * Default: 5m - * Group: consumer - * - * @param maxPollIntervalMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder maxPollIntervalMs( - int maxPollIntervalMs) { - doSetProperty("maxPollIntervalMs", maxPollIntervalMs); - return this; - } - /** - * The maximum delay between invocations of poll() when using consumer - * group management. This places an upper bound on the amount of time - * that the consumer can be idle before fetching more records. If poll() - * is not called before expiration of this timeout, then the consumer is - * considered failed and the group will rebalance in order to reassign - * the partitions to another member. For consumers using a non-null - * group.instance.id which reach this timeout, partitions will not be - * immediately reassigned. Instead, the consumer will stop sending - * heartbeats and partitions will be reassigned after expiration of - * session.timeout.ms. This mirrors the behavior of a static consumer - * which has shutdown. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 5m - * Group: consumer - * - * @param maxPollIntervalMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder maxPollIntervalMs( - String maxPollIntervalMs) { - doSetProperty("maxPollIntervalMs", maxPollIntervalMs); - return this; - } - /** - * The maximum number of records returned in a single call to poll(). - * Note, that max.poll.records does not impact the underlying fetching - * behavior. The consumer will cache the records from each fetch request - * and returns them incrementally from each poll. - * - * The option is a: <code>int</code> type. - * - * Default: 500 - * Group: consumer - * - * @param maxPollRecords the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder maxPollRecords( - int maxPollRecords) { - doSetProperty("maxPollRecords", maxPollRecords); - return this; - } - /** - * The maximum number of records returned in a single call to poll(). - * Note, that max.poll.records does not impact the underlying fetching - * behavior. The consumer will cache the records from each fetch request - * and returns them incrementally from each poll. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 500 - * Group: consumer - * - * @param maxPollRecords the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder maxPollRecords( - String maxPollRecords) { - doSetProperty("maxPollRecords", maxPollRecords); - return this; - } - /** - * A list of class names or class types, ordered by preference, of - * supported partition assignment strategies that the client will use to - * distribute partition ownership amongst consumer instances when group - * management is used. Available options - * are:org.apache.kafka.clients.consumer.RangeAssignor: The default - * assignor, which works on a per-topic - * basis.org.apache.kafka.clients.consumer.RoundRobinAssignor: Assigns - * partitions to consumers in a round-robin - * fashion.org.apache.kafka.clients.consumer.StickyAssignor: Guarantees - * an assignment that is maximally balanced while preserving as many - * existing partition assignments as - * possible.org.apache.kafka.clients.consumer.CooperativeStickyAssignor: - * Follows the same StickyAssignor logic, but allows for cooperative - * rebalancing.Implementing the - * org.apache.kafka.clients.consumer.ConsumerPartitionAssignor interface - * allows you to plug in a custom assignment strategy. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: org.apache.kafka.clients.consumer.RangeAssignor - * Group: consumer - * - * @param partitionAssignmentStrategy the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder partitionAssignmentStrategy( - String partitionAssignmentStrategy) { - doSetProperty("partitionAssignmentStrategy", partitionAssignmentStrategy); - return this; - } - /** - * Set if KafkaConsumer will read from a particular offset on startup. - * This config will take precedence over seekTo config. - * - * The option is a: <code>java.lang.Long</code> type. - * - * Group: consumer - * - * @param seekToOffset the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder seekToOffset(Long seekToOffset) { - doSetProperty("seekToOffset", seekToOffset); - return this; - } - /** - * Set if KafkaConsumer will read from a particular offset on startup. - * This config will take precedence over seekTo config. - * - * The option will be converted to a - * <code>java.lang.Long</code> type. - * - * Group: consumer - * - * @param seekToOffset the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder seekToOffset( - String seekToOffset) { - doSetProperty("seekToOffset", seekToOffset); - return this; - } - /** - * Set if KafkaConsumer will read from beginning or end on startup: - * beginning : read from beginning end : read from end. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: consumer - * - * @param seekToPosition the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder seekToPosition( - String seekToPosition) { - doSetProperty("seekToPosition", seekToPosition); - return this; - } - /** - * The timeout used to detect client failures when using Kafka's group - * management facility. The client sends periodic heartbeats to indicate - * its liveness to the broker. If no heartbeats are received by the - * broker before the expiration of this session timeout, then the broker - * will remove this client from the group and initiate a rebalance. Note - * that the value must be in the allowable range as configured in the - * broker configuration by group.min.session.timeout.ms and - * group.max.session.timeout.ms. - * - * The option is a: <code>int</code> type. - * - * Default: 10s - * Group: consumer - * - * @param sessionTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sessionTimeoutMs( - int sessionTimeoutMs) { - doSetProperty("sessionTimeoutMs", sessionTimeoutMs); - return this; - } - /** - * The timeout used to detect client failures when using Kafka's group - * management facility. The client sends periodic heartbeats to indicate - * its liveness to the broker. If no heartbeats are received by the - * broker before the expiration of this session timeout, then the broker - * will remove this client from the group and initiate a rebalance. Note - * that the value must be in the allowable range as configured in the - * broker configuration by group.min.session.timeout.ms and - * group.max.session.timeout.ms. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 10s - * Group: consumer - * - * @param sessionTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sessionTimeoutMs( - String sessionTimeoutMs) { - doSetProperty("sessionTimeoutMs", sessionTimeoutMs); - return this; - } - /** - * Deserializer class for value that implements the - * org.apache.kafka.common.serialization.Deserializer interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: org.apache.kafka.common.serialization.StringDeserializer - * Group: consumer - * - * @param valueDeserializer the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder valueDeserializer( - String valueDeserializer) { - doSetProperty("valueDeserializer", valueDeserializer); - return this; - } - /** - * The fully qualified name of a SASL client callback handler class that - * implements the AuthenticateCallbackHandler interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslClientCallbackHandlerClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslClientCallbackHandlerClass( - String saslClientCallbackHandlerClass) { - doSetProperty("saslClientCallbackHandlerClass", saslClientCallbackHandlerClass); - return this; - } - /** - * JAAS login context parameters for SASL connections in the format used - * by JAAS configuration files. JAAS configuration file format is - * described here. The format for the value is: loginModuleClass - * controlFlag (optionName=optionValue);. For brokers, the config must - * be prefixed with listener prefix and SASL mechanism name in - * lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=com.example.ScramLoginModule required;. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslJaasConfig the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslJaasConfig( - String saslJaasConfig) { - doSetProperty("saslJaasConfig", saslJaasConfig); - return this; - } - /** - * Kerberos kinit command path. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: /usr/bin/kinit - * Group: security - * - * @param saslKerberosKinitCmd the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslKerberosKinitCmd( - String saslKerberosKinitCmd) { - doSetProperty("saslKerberosKinitCmd", saslKerberosKinitCmd); - return this; - } - /** - * Login thread sleep time between refresh attempts. - * - * The option is a: <code>long</code> type. - * - * Default: 60000 - * Group: security - * - * @param saslKerberosMinTimeBeforeRelogin the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslKerberosMinTimeBeforeRelogin( - long saslKerberosMinTimeBeforeRelogin) { - doSetProperty("saslKerberosMinTimeBeforeRelogin", saslKerberosMinTimeBeforeRelogin); - return this; - } - /** - * Login thread sleep time between refresh attempts. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 60000 - * Group: security - * - * @param saslKerberosMinTimeBeforeRelogin the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslKerberosMinTimeBeforeRelogin( - String saslKerberosMinTimeBeforeRelogin) { - doSetProperty("saslKerberosMinTimeBeforeRelogin", saslKerberosMinTimeBeforeRelogin); - return this; - } - /** - * The Kerberos principal name that Kafka runs as. This can be defined - * either in Kafka's JAAS config or in Kafka's config. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslKerberosServiceName the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslKerberosServiceName( - String saslKerberosServiceName) { - doSetProperty("saslKerberosServiceName", saslKerberosServiceName); - return this; - } - /** - * Percentage of random jitter added to the renewal time. - * - * The option is a: <code>double</code> type. - * - * Default: 0.05 - * Group: security - * - * @param saslKerberosTicketRenewJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslKerberosTicketRenewJitter( - double saslKerberosTicketRenewJitter) { - doSetProperty("saslKerberosTicketRenewJitter", saslKerberosTicketRenewJitter); - return this; - } - /** - * Percentage of random jitter added to the renewal time. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.05 - * Group: security - * - * @param saslKerberosTicketRenewJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslKerberosTicketRenewJitter( - String saslKerberosTicketRenewJitter) { - doSetProperty("saslKerberosTicketRenewJitter", saslKerberosTicketRenewJitter); - return this; - } - /** - * Login thread will sleep until the specified window factor of time - * from last refresh to ticket's expiry has been reached, at which time - * it will try to renew the ticket. - * - * The option is a: <code>double</code> type. - * - * Default: 0.8 - * Group: security - * - * @param saslKerberosTicketRenewWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslKerberosTicketRenewWindowFactor( - double saslKerberosTicketRenewWindowFactor) { - doSetProperty("saslKerberosTicketRenewWindowFactor", saslKerberosTicketRenewWindowFactor); - return this; - } - /** - * Login thread will sleep until the specified window factor of time - * from last refresh to ticket's expiry has been reached, at which time - * it will try to renew the ticket. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.8 - * Group: security - * - * @param saslKerberosTicketRenewWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslKerberosTicketRenewWindowFactor( - String saslKerberosTicketRenewWindowFactor) { - doSetProperty("saslKerberosTicketRenewWindowFactor", saslKerberosTicketRenewWindowFactor); - return this; - } - /** - * The fully qualified name of a SASL login callback handler class that - * implements the AuthenticateCallbackHandler interface. For brokers, - * login callback handler config must be prefixed with listener prefix - * and SASL mechanism name in lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.login.callback.handler.class=com.example.CustomScramLoginCallbackHandler. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslLoginCallbackHandlerClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginCallbackHandlerClass( - String saslLoginCallbackHandlerClass) { - doSetProperty("saslLoginCallbackHandlerClass", saslLoginCallbackHandlerClass); - return this; - } - /** - * The fully qualified name of a class that implements the Login - * interface. For brokers, login config must be prefixed with listener - * prefix and SASL mechanism name in lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.login.class=com.example.CustomScramLogin. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslLoginClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginClass( - String saslLoginClass) { - doSetProperty("saslLoginClass", saslLoginClass); - return this; - } - /** - * The amount of buffer time before credential expiration to maintain - * when refreshing a credential, in seconds. If a refresh would - * otherwise occur closer to expiration than the number of buffer - * seconds then the refresh will be moved up to maintain as much of the - * buffer time as possible. Legal values are between 0 and 3600 (1 - * hour); a default value of 300 (5 minutes) is used if no value is - * specified. This value and sasl.login.refresh.min.period.seconds are - * both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>short</code> type. - * - * Default: 300 - * Group: security - * - * @param saslLoginRefreshBufferSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginRefreshBufferSeconds( - short saslLoginRefreshBufferSeconds) { - doSetProperty("saslLoginRefreshBufferSeconds", saslLoginRefreshBufferSeconds); - return this; - } - /** - * The amount of buffer time before credential expiration to maintain - * when refreshing a credential, in seconds. If a refresh would - * otherwise occur closer to expiration than the number of buffer - * seconds then the refresh will be moved up to maintain as much of the - * buffer time as possible. Legal values are between 0 and 3600 (1 - * hour); a default value of 300 (5 minutes) is used if no value is - * specified. This value and sasl.login.refresh.min.period.seconds are - * both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>short</code> - * type. - * - * Default: 300 - * Group: security - * - * @param saslLoginRefreshBufferSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginRefreshBufferSeconds( - String saslLoginRefreshBufferSeconds) { - doSetProperty("saslLoginRefreshBufferSeconds", saslLoginRefreshBufferSeconds); - return this; - } - /** - * The desired minimum time for the login refresh thread to wait before - * refreshing a credential, in seconds. Legal values are between 0 and - * 900 (15 minutes); a default value of 60 (1 minute) is used if no - * value is specified. This value and sasl.login.refresh.buffer.seconds - * are both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>short</code> type. - * - * Default: 60 - * Group: security - * - * @param saslLoginRefreshMinPeriodSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginRefreshMinPeriodSeconds( - short saslLoginRefreshMinPeriodSeconds) { - doSetProperty("saslLoginRefreshMinPeriodSeconds", saslLoginRefreshMinPeriodSeconds); - return this; - } - /** - * The desired minimum time for the login refresh thread to wait before - * refreshing a credential, in seconds. Legal values are between 0 and - * 900 (15 minutes); a default value of 60 (1 minute) is used if no - * value is specified. This value and sasl.login.refresh.buffer.seconds - * are both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>short</code> - * type. - * - * Default: 60 - * Group: security - * - * @param saslLoginRefreshMinPeriodSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginRefreshMinPeriodSeconds( - String saslLoginRefreshMinPeriodSeconds) { - doSetProperty("saslLoginRefreshMinPeriodSeconds", saslLoginRefreshMinPeriodSeconds); - return this; - } - /** - * Login refresh thread will sleep until the specified window factor - * relative to the credential's lifetime has been reached, at which time - * it will try to refresh the credential. Legal values are between 0.5 - * (50%) and 1.0 (100%) inclusive; a default value of 0.8 (80%) is used - * if no value is specified. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>double</code> type. - * - * Default: 0.8 - * Group: security - * - * @param saslLoginRefreshWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginRefreshWindowFactor( - double saslLoginRefreshWindowFactor) { - doSetProperty("saslLoginRefreshWindowFactor", saslLoginRefreshWindowFactor); - return this; - } - /** - * Login refresh thread will sleep until the specified window factor - * relative to the credential's lifetime has been reached, at which time - * it will try to refresh the credential. Legal values are between 0.5 - * (50%) and 1.0 (100%) inclusive; a default value of 0.8 (80%) is used - * if no value is specified. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.8 - * Group: security - * - * @param saslLoginRefreshWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginRefreshWindowFactor( - String saslLoginRefreshWindowFactor) { - doSetProperty("saslLoginRefreshWindowFactor", saslLoginRefreshWindowFactor); - return this; - } - /** - * The maximum amount of random jitter relative to the credential's - * lifetime that is added to the login refresh thread's sleep time. - * Legal values are between 0 and 0.25 (25%) inclusive; a default value - * of 0.05 (5%) is used if no value is specified. Currently applies only - * to OAUTHBEARER. - * - * The option is a: <code>double</code> type. - * - * Default: 0.05 - * Group: security - * - * @param saslLoginRefreshWindowJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginRefreshWindowJitter( - double saslLoginRefreshWindowJitter) { - doSetProperty("saslLoginRefreshWindowJitter", saslLoginRefreshWindowJitter); - return this; - } - /** - * The maximum amount of random jitter relative to the credential's - * lifetime that is added to the login refresh thread's sleep time. - * Legal values are between 0 and 0.25 (25%) inclusive; a default value - * of 0.05 (5%) is used if no value is specified. Currently applies only - * to OAUTHBEARER. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.05 - * Group: security - * - * @param saslLoginRefreshWindowJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslLoginRefreshWindowJitter( - String saslLoginRefreshWindowJitter) { - doSetProperty("saslLoginRefreshWindowJitter", saslLoginRefreshWindowJitter); - return this; - } - /** - * SASL mechanism used for client connections. This may be any mechanism - * for which a security provider is available. GSSAPI is the default - * mechanism. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: GSSAPI - * Group: security - * - * @param saslMechanism the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder saslMechanism( - String saslMechanism) { - doSetProperty("saslMechanism", saslMechanism); - return this; - } - /** - * Protocol used to communicate with brokers. Valid values are: - * PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: PLAINTEXT - * Group: security - * - * @param securityProtocol the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder securityProtocol( - String securityProtocol) { - doSetProperty("securityProtocol", securityProtocol); - return this; - } - /** - * A list of configurable creator classes each returning a provider - * implementing security algorithms. These classes should implement the - * org.apache.kafka.common.security.auth.SecurityProviderCreator - * interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param securityProviders the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder securityProviders( - String securityProviders) { - doSetProperty("securityProviders", securityProviders); - return this; - } - /** - * A list of cipher suites. This is a named combination of - * authentication, encryption, MAC and key exchange algorithm used to - * negotiate the security settings for a network connection using TLS or - * SSL network protocol. By default all the available cipher suites are - * supported. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslCipherSuites the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslCipherSuites( - String sslCipherSuites) { - doSetProperty("sslCipherSuites", sslCipherSuites); - return this; - } - /** - * The list of protocols enabled for SSL connections. The default is - * 'TLSv1.2,TLSv1.3' when running with Java 11 or newer, 'TLSv1.2' - * otherwise. With the default value for Java 11, clients and servers - * will prefer TLSv1.3 if both support it and fallback to TLSv1.2 - * otherwise (assuming both support at least TLSv1.2). This default - * should be fine for most cases. Also see the config documentation for - * ssl.protocol. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: TLSv1.2,TLSv1.3 - * Group: security - * - * @param sslEnabledProtocols the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslEnabledProtocols( - String sslEnabledProtocols) { - doSetProperty("sslEnabledProtocols", sslEnabledProtocols); - return this; - } - /** - * The endpoint identification algorithm to validate server hostname - * using server certificate. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: https - * Group: security - * - * @param sslEndpointIdentificationAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslEndpointIdentificationAlgorithm( - String sslEndpointIdentificationAlgorithm) { - doSetProperty("sslEndpointIdentificationAlgorithm", sslEndpointIdentificationAlgorithm); - return this; - } - /** - * The class of type - * org.apache.kafka.common.security.auth.SslEngineFactory to provide - * SSLEngine objects. Default value is - * org.apache.kafka.common.security.ssl.DefaultSslEngineFactory. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslEngineFactoryClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslEngineFactoryClass( - String sslEngineFactoryClass) { - doSetProperty("sslEngineFactoryClass", sslEngineFactoryClass); - return this; - } - /** - * The algorithm used by key manager factory for SSL connections. - * Default value is the key manager factory algorithm configured for the - * Java Virtual Machine. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: SunX509 - * Group: security - * - * @param sslKeymanagerAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslKeymanagerAlgorithm( - String sslKeymanagerAlgorithm) { - doSetProperty("sslKeymanagerAlgorithm", sslKeymanagerAlgorithm); - return this; - } - /** - * The password of the private key in the key store file orthe PEM key - * specified in ssl.keystore.key'. This is required for clients only if - * two-way authentication is configured. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeyPassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslKeyPassword( - String sslKeyPassword) { - doSetProperty("sslKeyPassword", sslKeyPassword); - return this; - } - /** - * Certificate chain in the format specified by 'ssl.keystore.type'. - * Default SSL engine factory supports only PEM format with a list of - * X.509 certificates. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreCertificateChain the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslKeystoreCertificateChain( - String sslKeystoreCertificateChain) { - doSetProperty("sslKeystoreCertificateChain", sslKeystoreCertificateChain); - return this; - } - /** - * Private key in the format specified by 'ssl.keystore.type'. Default - * SSL engine factory supports only PEM format with PKCS#8 keys. If the - * key is encrypted, key password must be specified using - * 'ssl.key.password'. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreKey the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslKeystoreKey( - String sslKeystoreKey) { - doSetProperty("sslKeystoreKey", sslKeystoreKey); - return this; - } - /** - * The location of the key store file. This is optional for client and - * can be used for two-way authentication for client. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreLocation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslKeystoreLocation( - String sslKeystoreLocation) { - doSetProperty("sslKeystoreLocation", sslKeystoreLocation); - return this; - } - /** - * The store password for the key store file. This is optional for - * client and only needed if 'ssl.keystore.location' is configured. Key - * store password is not supported for PEM format. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystorePassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslKeystorePassword( - String sslKeystorePassword) { - doSetProperty("sslKeystorePassword", sslKeystorePassword); - return this; - } - /** - * The file format of the key store file. This is optional for client. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: JKS - * Group: security - * - * @param sslKeystoreType the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslKeystoreType( - String sslKeystoreType) { - doSetProperty("sslKeystoreType", sslKeystoreType); - return this; - } - /** - * The SSL protocol used to generate the SSLContext. The default is - * 'TLSv1.3' when running with Java 11 or newer, 'TLSv1.2' otherwise. - * This value should be fine for most use cases. Allowed values in - * recent JVMs are 'TLSv1.2' and 'TLSv1.3'. 'TLS', 'TLSv1.1', 'SSL', - * 'SSLv2' and 'SSLv3' may be supported in older JVMs, but their usage - * is discouraged due to known security vulnerabilities. With the - * default value for this config and 'ssl.enabled.protocols', clients - * will downgrade to 'TLSv1.2' if the server does not support 'TLSv1.3'. - * If this config is set to 'TLSv1.2', clients will not use 'TLSv1.3' - * even if it is one of the values in ssl.enabled.protocols and the - * server only supports 'TLSv1.3'. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: TLSv1.2 - * Group: security - * - * @param sslProtocol the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslProtocol(String sslProtocol) { - doSetProperty("sslProtocol", sslProtocol); - return this; - } - /** - * The name of the security provider used for SSL connections. Default - * value is the default security provider of the JVM. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslProvider the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslProvider(String sslProvider) { - doSetProperty("sslProvider", sslProvider); - return this; - } - /** - * The SecureRandom PRNG implementation to use for SSL cryptography - * operations. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslSecureRandomImplementation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslSecureRandomImplementation( - String sslSecureRandomImplementation) { - doSetProperty("sslSecureRandomImplementation", sslSecureRandomImplementation); - return this; - } - /** - * The algorithm used by trust manager factory for SSL connections. - * Default value is the trust manager factory algorithm configured for - * the Java Virtual Machine. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: PKIX - * Group: security - * - * @param sslTrustmanagerAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslTrustmanagerAlgorithm( - String sslTrustmanagerAlgorithm) { - doSetProperty("sslTrustmanagerAlgorithm", sslTrustmanagerAlgorithm); - return this; - } - /** - * Trusted certificates in the format specified by - * 'ssl.truststore.type'. Default SSL engine factory supports only PEM - * format with X.509 certificates. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststoreCertificates the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslTruststoreCertificates( - String sslTruststoreCertificates) { - doSetProperty("sslTruststoreCertificates", sslTruststoreCertificates); - return this; - } - /** - * The location of the trust store file. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststoreLocation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslTruststoreLocation( - String sslTruststoreLocation) { - doSetProperty("sslTruststoreLocation", sslTruststoreLocation); - return this; - } - /** - * The password for the trust store file. If a password is not set, - * trust store file configured will still be used, but integrity - * checking is disabled. Trust store password is not supported for PEM - * format. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststorePassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslTruststorePassword( - String sslTruststorePassword) { - doSetProperty("sslTruststorePassword", sslTruststorePassword); - return this; - } - /** - * The file format of the trust store file. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: JKS - * Group: security - * - * @param sslTruststoreType the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointConsumerBuilder sslTruststoreType( - String sslTruststoreType) { - doSetProperty("sslTruststoreType", sslTruststoreType); - return this; - } - } - - /** - * Advanced builder for endpoint consumers for the Vert.x Kafka component. - */ - public interface AdvancedVertxKafkaEndpointConsumerBuilder - extends - EndpointConsumerBuilder { - default VertxKafkaEndpointConsumerBuilder basic() { - return (VertxKafkaEndpointConsumerBuilder) this; - } - /** - * Allows for bridging the consumer to the Camel routing Error Handler, - * which mean any exceptions occurred while the consumer is trying to - * pickup incoming messages, or the likes, will now be processed as a - * message and handled by the routing Error Handler. By default the - * consumer will use the org.apache.camel.spi.ExceptionHandler to deal - * with exceptions, that will be logged at WARN or ERROR level and - * ignored. - * - * The option is a: <code>boolean</code> type. - * - * Default: false - * Group: consumer (advanced) - * - * @param bridgeErrorHandler the value to set - * @return the dsl builder - */ - default AdvancedVertxKafkaEndpointConsumerBuilder bridgeErrorHandler( - boolean bridgeErrorHandler) { - doSetProperty("bridgeErrorHandler", bridgeErrorHandler); - return this; - } - /** - * Allows for bridging the consumer to the Camel routing Error Handler, - * which mean any exceptions occurred while the consumer is trying to - * pickup incoming messages, or the likes, will now be processed as a - * message and handled by the routing Error Handler. By default the - * consumer will use the org.apache.camel.spi.ExceptionHandler to deal - * with exceptions, that will be logged at WARN or ERROR level and - * ignored. - * - * The option will be converted to a <code>boolean</code> - * type. - * - * Default: false - * Group: consumer (advanced) - * - * @param bridgeErrorHandler the value to set - * @return the dsl builder - */ - default AdvancedVertxKafkaEndpointConsumerBuilder bridgeErrorHandler( - String bridgeErrorHandler) { - doSetProperty("bridgeErrorHandler", bridgeErrorHandler); - return this; - } - /** - * To let the consumer use a custom ExceptionHandler. Notice if the - * option bridgeErrorHandler is enabled then this option is not in use. - * By default the consumer will deal with exceptions, that will be - * logged at WARN or ERROR level and ignored. - * - * The option is a: - * <code>org.apache.camel.spi.ExceptionHandler</code> type. - * - * Group: consumer (advanced) - * - * @param exceptionHandler the value to set - * @return the dsl builder - */ - default AdvancedVertxKafkaEndpointConsumerBuilder exceptionHandler( - org.apache.camel.spi.ExceptionHandler exceptionHandler) { - doSetProperty("exceptionHandler", exceptionHandler); - return this; - } - /** - * To let the consumer use a custom ExceptionHandler. Notice if the - * option bridgeErrorHandler is enabled then this option is not in use. - * By default the consumer will deal with exceptions, that will be - * logged at WARN or ERROR level and ignored. - * - * The option will be converted to a - * <code>org.apache.camel.spi.ExceptionHandler</code> type. - * - * Group: consumer (advanced) - * - * @param exceptionHandler the value to set - * @return the dsl builder - */ - default AdvancedVertxKafkaEndpointConsumerBuilder exceptionHandler( - String exceptionHandler) { - doSetProperty("exceptionHandler", exceptionHandler); - return this; - } - /** - * Sets the exchange pattern when the consumer creates an exchange. - * - * The option is a: - * <code>org.apache.camel.ExchangePattern</code> type. - * - * Group: consumer (advanced) - * - * @param exchangePattern the value to set - * @return the dsl builder - */ - default AdvancedVertxKafkaEndpointConsumerBuilder exchangePattern( - org.apache.camel.ExchangePattern exchangePattern) { - doSetProperty("exchangePattern", exchangePattern); - return this; - } - /** - * Sets the exchange pattern when the consumer creates an exchange. - * - * The option will be converted to a - * <code>org.apache.camel.ExchangePattern</code> type. - * - * Group: consumer (advanced) - * - * @param exchangePattern the value to set - * @return the dsl builder - */ - default AdvancedVertxKafkaEndpointConsumerBuilder exchangePattern( - String exchangePattern) { - doSetProperty("exchangePattern", exchangePattern); - return this; - } - } - - /** - * Builder for endpoint producers for the Vert.x Kafka component. - */ - public interface VertxKafkaEndpointProducerBuilder - extends - EndpointProducerBuilder { - default AdvancedVertxKafkaEndpointProducerBuilder advanced() { - return (AdvancedVertxKafkaEndpointProducerBuilder) this; - } - /** - * Sets additional properties for either kafka consumer or kafka - * producer in case they can't be set directly on the camel - * configurations (e.g: new Kafka properties that are not reflected yet - * in Camel configurations), the properties have to be prefixed with - * additionalProperties.. E.g: - * additionalProperties.transactional.id=12345&amp;additionalProperties.schema.registry.url=http://localhost:8811/avro. - * - * The option is a: <code>java.util.Map&lt;java.lang.String, - * java.lang.Object&gt;</code> type. - * The option is multivalued, and you can use the - * additionalProperties(String, Object) method to add a value (call the - * method multiple times to set more values). - * - * Group: common - * - * @param key the option key - * @param value the option value - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder additionalProperties( - String key, - Object value) { - doSetMultiValueProperty("additionalProperties", "additionalProperties." + key, value); - return this; - } - /** - * Sets additional properties for either kafka consumer or kafka - * producer in case they can't be set directly on the camel - * configurations (e.g: new Kafka properties that are not reflected yet - * in Camel configurations), the properties have to be prefixed with - * additionalProperties.. E.g: - * additionalProperties.transactional.id=12345&amp;additionalProperties.schema.registry.url=http://localhost:8811/avro. - * - * The option is a: <code>java.util.Map&lt;java.lang.String, - * java.lang.Object&gt;</code> type. - * The option is multivalued, and you can use the - * additionalProperties(String, Object) method to add a value (call the - * method multiple times to set more values). - * - * Group: common - * - * @param values the values - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder additionalProperties( - Map values) { - doSetMultiValueProperties("additionalProperties", "additionalProperties.", values); - return this; - } - /** - * A list of host/port pairs to use for establishing the initial - * connection to the Kafka cluster. The client will make use of all - * servers irrespective of which servers are specified here for - * bootstrapping&amp;mdash;this list only impacts the initial hosts - * used to discover the full set of servers. This list should be in the - * form host1:port1,host2:port2,.... Since these servers are just used - * for the initial connection to discover the full cluster membership - * (which may change dynamically), this list need not contain the full - * set of servers (you may want more than one, though, in case a server - * is down). - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param bootstrapServers the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder bootstrapServers( - String bootstrapServers) { - doSetProperty("bootstrapServers", bootstrapServers); - return this; - } - /** - * Controls how the client uses DNS lookups. If set to use_all_dns_ips, - * connect to each returned IP address in sequence until a successful - * connection is established. After a disconnection, the next IP is - * used. Once all IPs have been used once, the client resolves the IP(s) - * from the hostname again (both the JVM and the OS cache DNS name - * lookups, however). If set to - * resolve_canonical_bootstrap_servers_only, resolve each bootstrap - * address into a list of canonical names. After the bootstrap phase, - * this behaves the same as use_all_dns_ips. If set to default - * (deprecated), attempt to connect to the first IP address returned by - * the lookup, even if the lookup returns multiple IP addresses. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: use_all_dns_ips - * Group: common - * - * @param clientDnsLookup the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder clientDnsLookup( - String clientDnsLookup) { - doSetProperty("clientDnsLookup", clientDnsLookup); - return this; - } - /** - * An id string to pass to the server when making requests. The purpose - * of this is to be able to track the source of requests beyond just - * ip/port by allowing a logical application name to be included in - * server-side request logging. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param clientId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder clientId(String clientId) { - doSetProperty("clientId", clientId); - return this; - } - /** - * Close idle connections after the number of milliseconds specified by - * this config. - * - * The option is a: <code>long</code> type. - * - * Default: 9m - * Group: common - * - * @param connectionsMaxIdleMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder connectionsMaxIdleMs( - long connectionsMaxIdleMs) { - doSetProperty("connectionsMaxIdleMs", connectionsMaxIdleMs); - return this; - } - /** - * Close idle connections after the number of milliseconds specified by - * this config. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 9m - * Group: common - * - * @param connectionsMaxIdleMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder connectionsMaxIdleMs( - String connectionsMaxIdleMs) { - doSetProperty("connectionsMaxIdleMs", connectionsMaxIdleMs); - return this; - } - /** - * To use a custom HeaderFilterStrategy to filter header to and from - * Camel message. - * - * The option is a: - * <code>org.apache.camel.spi.HeaderFilterStrategy</code> - * type. - * - * Group: common - * - * @param headerFilterStrategy the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder headerFilterStrategy( - org.apache.camel.spi.HeaderFilterStrategy headerFilterStrategy) { - doSetProperty("headerFilterStrategy", headerFilterStrategy); - return this; - } - /** - * To use a custom HeaderFilterStrategy to filter header to and from - * Camel message. - * - * The option will be converted to a - * <code>org.apache.camel.spi.HeaderFilterStrategy</code> - * type. - * - * Group: common - * - * @param headerFilterStrategy the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder headerFilterStrategy( - String headerFilterStrategy) { - doSetProperty("headerFilterStrategy", headerFilterStrategy); - return this; - } - /** - * A list of classes to use as interceptors. Implementing the - * org.apache.kafka.clients.producer.ProducerInterceptor interface - * allows you to intercept (and possibly mutate) the records received by - * the producer before they are published to the Kafka cluster. By - * default, there are no interceptors. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param interceptorClasses the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder interceptorClasses( - String interceptorClasses) { - doSetProperty("interceptorClasses", interceptorClasses); - return this; - } - /** - * The period of time in milliseconds after which we force a refresh of - * metadata even if we haven't seen any partition leadership changes to - * proactively discover any new brokers or partitions. - * - * The option is a: <code>long</code> type. - * - * Default: 5m - * Group: common - * - * @param metadataMaxAgeMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metadataMaxAgeMs( - long metadataMaxAgeMs) { - doSetProperty("metadataMaxAgeMs", metadataMaxAgeMs); - return this; - } - /** - * The period of time in milliseconds after which we force a refresh of - * metadata even if we haven't seen any partition leadership changes to - * proactively discover any new brokers or partitions. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 5m - * Group: common - * - * @param metadataMaxAgeMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metadataMaxAgeMs( - String metadataMaxAgeMs) { - doSetProperty("metadataMaxAgeMs", metadataMaxAgeMs); - return this; - } - /** - * A list of classes to use as metrics reporters. Implementing the - * org.apache.kafka.common.metrics.MetricsReporter interface allows - * plugging in classes that will be notified of new metric creation. The - * JmxReporter is always included to register JMX statistics. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param metricReporters the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metricReporters( - String metricReporters) { - doSetProperty("metricReporters", metricReporters); - return this; - } - /** - * The number of samples maintained to compute metrics. - * - * The option is a: <code>int</code> type. - * - * Default: 2 - * Group: common - * - * @param metricsNumSamples the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metricsNumSamples( - int metricsNumSamples) { - doSetProperty("metricsNumSamples", metricsNumSamples); - return this; - } - /** - * The number of samples maintained to compute metrics. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 2 - * Group: common - * - * @param metricsNumSamples the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metricsNumSamples( - String metricsNumSamples) { - doSetProperty("metricsNumSamples", metricsNumSamples); - return this; - } - /** - * The highest recording level for metrics. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: INFO - * Group: common - * - * @param metricsRecordingLevel the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metricsRecordingLevel( - String metricsRecordingLevel) { - doSetProperty("metricsRecordingLevel", metricsRecordingLevel); - return this; - } - /** - * The window of time a metrics sample is computed over. - * - * The option is a: <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param metricsSampleWindowMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metricsSampleWindowMs( - long metricsSampleWindowMs) { - doSetProperty("metricsSampleWindowMs", metricsSampleWindowMs); - return this; - } - /** - * The window of time a metrics sample is computed over. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param metricsSampleWindowMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metricsSampleWindowMs( - String metricsSampleWindowMs) { - doSetProperty("metricsSampleWindowMs", metricsSampleWindowMs); - return this; - } - /** - * The partition to which the record will be sent (or null if no - * partition was specified) or read from a particular partition if set. - * Header VertxKafkaConstants#PARTITION_ID If configured, it will take - * precedence over this config. - * - * The option is a: <code>java.lang.Integer</code> type. - * - * Group: common - * - * @param partitionId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder partitionId( - Integer partitionId) { - doSetProperty("partitionId", partitionId); - return this; - } - /** - * The partition to which the record will be sent (or null if no - * partition was specified) or read from a particular partition if set. - * Header VertxKafkaConstants#PARTITION_ID If configured, it will take - * precedence over this config. - * - * The option will be converted to a - * <code>java.lang.Integer</code> type. - * - * Group: common - * - * @param partitionId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder partitionId(String partitionId) { - doSetProperty("partitionId", partitionId); - return this; - } - /** - * The size of the TCP receive buffer (SO_RCVBUF) to use when reading - * data. If the value is -1, the OS default will be used. - * - * The option is a: <code>int</code> type. - * - * Default: 32768 - * Group: common - * - * @param receiveBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder receiveBufferBytes( - int receiveBufferBytes) { - doSetProperty("receiveBufferBytes", receiveBufferBytes); - return this; - } - /** - * The size of the TCP receive buffer (SO_RCVBUF) to use when reading - * data. If the value is -1, the OS default will be used. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 32768 - * Group: common - * - * @param receiveBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder receiveBufferBytes( - String receiveBufferBytes) { - doSetProperty("receiveBufferBytes", receiveBufferBytes); - return this; - } - /** - * The maximum amount of time in milliseconds to wait when reconnecting - * to a broker that has repeatedly failed to connect. If provided, the - * backoff per host will increase exponentially for each consecutive - * connection failure, up to this maximum. After calculating the backoff - * increase, 20% random jitter is added to avoid connection storms. - * - * The option is a: <code>long</code> type. - * - * Default: 1s - * Group: common - * - * @param reconnectBackoffMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder reconnectBackoffMaxMs( - long reconnectBackoffMaxMs) { - doSetProperty("reconnectBackoffMaxMs", reconnectBackoffMaxMs); - return this; - } - /** - * The maximum amount of time in milliseconds to wait when reconnecting - * to a broker that has repeatedly failed to connect. If provided, the - * backoff per host will increase exponentially for each consecutive - * connection failure, up to this maximum. After calculating the backoff - * increase, 20% random jitter is added to avoid connection storms. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 1s - * Group: common - * - * @param reconnectBackoffMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder reconnectBackoffMaxMs( - String reconnectBackoffMaxMs) { - doSetProperty("reconnectBackoffMaxMs", reconnectBackoffMaxMs); - return this; - } - /** - * The base amount of time to wait before attempting to reconnect to a - * given host. This avoids repeatedly connecting to a host in a tight - * loop. This backoff applies to all connection attempts by the client - * to a broker. - * - * The option is a: <code>long</code> type. - * - * Default: 50ms - * Group: common - * - * @param reconnectBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder reconnectBackoffMs( - long reconnectBackoffMs) { - doSetProperty("reconnectBackoffMs", reconnectBackoffMs); - return this; - } - /** - * The base amount of time to wait before attempting to reconnect to a - * given host. This avoids repeatedly connecting to a host in a tight - * loop. This backoff applies to all connection attempts by the client - * to a broker. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 50ms - * Group: common - * - * @param reconnectBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder reconnectBackoffMs( - String reconnectBackoffMs) { - doSetProperty("reconnectBackoffMs", reconnectBackoffMs); - return this; - } - /** - * The configuration controls the maximum amount of time the client will - * wait for the response of a request. If the response is not received - * before the timeout elapses the client will resend the request if - * necessary or fail the request if retries are exhausted. This should - * be larger than replica.lag.time.max.ms (a broker configuration) to - * reduce the possibility of message duplication due to unnecessary - * producer retries. - * - * The option is a: <code>int</code> type. - * - * Default: 30s - * Group: common - * - * @param requestTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder requestTimeoutMs( - int requestTimeoutMs) { - doSetProperty("requestTimeoutMs", requestTimeoutMs); - return this; - } - /** - * The configuration controls the maximum amount of time the client will - * wait for the response of a request. If the response is not received - * before the timeout elapses the client will resend the request if - * necessary or fail the request if retries are exhausted. This should - * be larger than replica.lag.time.max.ms (a broker configuration) to - * reduce the possibility of message duplication due to unnecessary - * producer retries. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 30s - * Group: common - * - * @param requestTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder requestTimeoutMs( - String requestTimeoutMs) { - doSetProperty("requestTimeoutMs", requestTimeoutMs); - return this; - } - /** - * The amount of time to wait before attempting to retry a failed - * request to a given topic partition. This avoids repeatedly sending - * requests in a tight loop under some failure scenarios. - * - * The option is a: <code>long</code> type. - * - * Default: 100ms - * Group: common - * - * @param retryBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder retryBackoffMs( - long retryBackoffMs) { - doSetProperty("retryBackoffMs", retryBackoffMs); - return this; - } - /** - * The amount of time to wait before attempting to retry a failed - * request to a given topic partition. This avoids repeatedly sending - * requests in a tight loop under some failure scenarios. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 100ms - * Group: common - * - * @param retryBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder retryBackoffMs( - String retryBackoffMs) { - doSetProperty("retryBackoffMs", retryBackoffMs); - return this; - } - /** - * The size of the TCP send buffer (SO_SNDBUF) to use when sending data. - * If the value is -1, the OS default will be used. - * - * The option is a: <code>int</code> type. - * - * Default: 131072 - * Group: common - * - * @param sendBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sendBufferBytes( - int sendBufferBytes) { - doSetProperty("sendBufferBytes", sendBufferBytes); - return this; - } - /** - * The size of the TCP send buffer (SO_SNDBUF) to use when sending data. - * If the value is -1, the OS default will be used. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 131072 - * Group: common - * - * @param sendBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sendBufferBytes( - String sendBufferBytes) { - doSetProperty("sendBufferBytes", sendBufferBytes); - return this; - } - /** - * The maximum amount of time the client will wait for the socket - * connection to be established. The connection setup timeout will - * increase exponentially for each consecutive connection failure up to - * this maximum. To avoid connection storms, a randomization factor of - * 0.2 will be applied to the timeout resulting in a random range - * between 20% below and 20% above the computed value. - * - * The option is a: <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param socketConnectionSetupTimeoutMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder socketConnectionSetupTimeoutMaxMs( - long socketConnectionSetupTimeoutMaxMs) { - doSetProperty("socketConnectionSetupTimeoutMaxMs", socketConnectionSetupTimeoutMaxMs); - return this; - } - /** - * The maximum amount of time the client will wait for the socket - * connection to be established. The connection setup timeout will - * increase exponentially for each consecutive connection failure up to - * this maximum. To avoid connection storms, a randomization factor of - * 0.2 will be applied to the timeout resulting in a random range - * between 20% below and 20% above the computed value. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param socketConnectionSetupTimeoutMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder socketConnectionSetupTimeoutMaxMs( - String socketConnectionSetupTimeoutMaxMs) { - doSetProperty("socketConnectionSetupTimeoutMaxMs", socketConnectionSetupTimeoutMaxMs); - return this; - } - /** - * The amount of time the client will wait for the socket connection to - * be established. If the connection is not built before the timeout - * elapses, clients will close the socket channel. - * - * The option is a: <code>long</code> type. - * - * Default: 10s - * Group: common - * - * @param socketConnectionSetupTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder socketConnectionSetupTimeoutMs( - long socketConnectionSetupTimeoutMs) { - doSetProperty("socketConnectionSetupTimeoutMs", socketConnectionSetupTimeoutMs); - return this; - } - /** - * The amount of time the client will wait for the socket connection to - * be established. If the connection is not built before the timeout - * elapses, clients will close the socket channel. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 10s - * Group: common - * - * @param socketConnectionSetupTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder socketConnectionSetupTimeoutMs( - String socketConnectionSetupTimeoutMs) { - doSetProperty("socketConnectionSetupTimeoutMs", socketConnectionSetupTimeoutMs); - return this; - } - /** - * The number of acknowledgments the producer requires the leader to - * have received before considering a request complete. This controls - * the durability of records that are sent. The following settings are - * allowed: acks=0 If set to zero then the producer will not wait for - * any acknowledgment from the server at all. The record will be - * immediately added to the socket buffer and considered sent. No - * guarantee can be made that the server has received the record in this - * case, and the retries configuration will not take effect (as the - * client won't generally know of any failures). The offset given back - * for each record will always be set to -1. acks=1 This will mean the - * leader will write the record to its local log but will respond - * without awaiting full acknowledgement from all followers. In this - * case should the leader fail immediately after acknowledging the - * record but before the followers have replicated it then the record - * will be lost. acks=all This means the leader will wait for the full - * set of in-sync replicas to acknowledge the record. This guarantees - * that the record will not be lost as long as at least one in-sync - * replica remains alive. This is the strongest available guarantee. - * This is equivalent to the acks=-1 setting. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: 1 - * Group: producer - * - * @param acks the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder acks(String acks) { - doSetProperty("acks", acks); - return this; - } - /** - * The producer will attempt to batch records together into fewer - * requests whenever multiple records are being sent to the same - * partition. This helps performance on both the client and the server. - * This configuration controls the default batch size in bytes. No - * attempt will be made to batch records larger than this size. Requests - * sent to brokers will contain multiple batches, one for each partition - * with data available to be sent. A small batch size will make batching - * less common and may reduce throughput (a batch size of zero will - * disable batching entirely). A very large batch size may use memory a - * bit more wastefully as we will always allocate a buffer of the - * specified batch size in anticipation of additional records. - * - * The option is a: <code>int</code> type. - * - * Default: 16384 - * Group: producer - * - * @param batchSize the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder batchSize(int batchSize) { - doSetProperty("batchSize", batchSize); - return this; - } - /** - * The producer will attempt to batch records together into fewer - * requests whenever multiple records are being sent to the same - * partition. This helps performance on both the client and the server. - * This configuration controls the default batch size in bytes. No - * attempt will be made to batch records larger than this size. Requests - * sent to brokers will contain multiple batches, one for each partition - * with data available to be sent. A small batch size will make batching - * less common and may reduce throughput (a batch size of zero will - * disable batching entirely). A very large batch size may use memory a - * bit more wastefully as we will always allocate a buffer of the - * specified batch size in anticipation of additional records. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 16384 - * Group: producer - * - * @param batchSize the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder batchSize(String batchSize) { - doSetProperty("batchSize", batchSize); - return this; - } - /** - * The total bytes of memory the producer can use to buffer records - * waiting to be sent to the server. If records are sent faster than - * they can be delivered to the server the producer will block for - * max.block.ms after which it will throw an exception.This setting - * should correspond roughly to the total memory the producer will use, - * but is not a hard bound since not all memory the producer uses is - * used for buffering. Some additional memory will be used for - * compression (if compression is enabled) as well as for maintaining - * in-flight requests. - * - * The option is a: <code>long</code> type. - * - * Default: 33554432 - * Group: producer - * - * @param bufferMemory the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder bufferMemory(long bufferMemory) { - doSetProperty("bufferMemory", bufferMemory); - return this; - } - /** - * The total bytes of memory the producer can use to buffer records - * waiting to be sent to the server. If records are sent faster than - * they can be delivered to the server the producer will block for - * max.block.ms after which it will throw an exception.This setting - * should correspond roughly to the total memory the producer will use, - * but is not a hard bound since not all memory the producer uses is - * used for buffering. Some additional memory will be used for - * compression (if compression is enabled) as well as for maintaining - * in-flight requests. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 33554432 - * Group: producer - * - * @param bufferMemory the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder bufferMemory( - String bufferMemory) { - doSetProperty("bufferMemory", bufferMemory); - return this; - } - /** - * The compression type for all data generated by the producer. The - * default is none (i.e. no compression). Valid values are none, gzip, - * snappy, lz4, or zstd. Compression is of full batches of data, so the - * efficacy of batching will also impact the compression ratio (more - * batching means better compression). - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: none - * Group: producer - * - * @param compressionType the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder compressionType( - String compressionType) { - doSetProperty("compressionType", compressionType); - return this; - } - /** - * An upper bound on the time to report success or failure after a call - * to send() returns. This limits the total time that a record will be - * delayed prior to sending, the time to await acknowledgement from the - * broker (if expected), and the time allowed for retriable send - * failures. The producer may report failure to send a record earlier - * than this config if either an unrecoverable error is encountered, the - * retries have been exhausted, or the record is added to a batch which - * reached an earlier delivery expiration deadline. The value of this - * config should be greater than or equal to the sum of - * request.timeout.ms and linger.ms. - * - * The option is a: <code>int</code> type. - * - * Default: 2m - * Group: producer - * - * @param deliveryTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder deliveryTimeoutMs( - int deliveryTimeoutMs) { - doSetProperty("deliveryTimeoutMs", deliveryTimeoutMs); - return this; - } - /** - * An upper bound on the time to report success or failure after a call - * to send() returns. This limits the total time that a record will be - * delayed prior to sending, the time to await acknowledgement from the - * broker (if expected), and the time allowed for retriable send - * failures. The producer may report failure to send a record earlier - * than this config if either an unrecoverable error is encountered, the - * retries have been exhausted, or the record is added to a batch which - * reached an earlier delivery expiration deadline. The value of this - * config should be greater than or equal to the sum of - * request.timeout.ms and linger.ms. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 2m - * Group: producer - * - * @param deliveryTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder deliveryTimeoutMs( - String deliveryTimeoutMs) { - doSetProperty("deliveryTimeoutMs", deliveryTimeoutMs); - return this; - } - /** - * When set to 'true', the producer will ensure that exactly one copy of - * each message is written in the stream. If 'false', producer retries - * due to broker failures, etc., may write duplicates of the retried - * message in the stream. Note that enabling idempotence requires - * max.in.flight.requests.per.connection to be less than or equal to 5, - * retries to be greater than 0 and acks must be 'all'. If these values - * are not explicitly set by the user, suitable values will be chosen. - * If incompatible values are set, a ConfigException will be thrown. - * - * The option is a: <code>boolean</code> type. - * - * Default: false - * Group: producer - * - * @param enableIdempotence the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder enableIdempotence( - boolean enableIdempotence) { - doSetProperty("enableIdempotence", enableIdempotence); - return this; - } - /** - * When set to 'true', the producer will ensure that exactly one copy of - * each message is written in the stream. If 'false', producer retries - * due to broker failures, etc., may write duplicates of the retried - * message in the stream. Note that enabling idempotence requires - * max.in.flight.requests.per.connection to be less than or equal to 5, - * retries to be greater than 0 and acks must be 'all'. If these values - * are not explicitly set by the user, suitable values will be chosen. - * If incompatible values are set, a ConfigException will be thrown. - * - * The option will be converted to a <code>boolean</code> - * type. - * - * Default: false - * Group: producer - * - * @param enableIdempotence the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder enableIdempotence( - String enableIdempotence) { - doSetProperty("enableIdempotence", enableIdempotence); - return this; - } - /** - * Serializer class for key that implements the - * org.apache.kafka.common.serialization.Serializer interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: org.apache.kafka.common.serialization.StringSerializer - * Group: producer - * - * @param keySerializer the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder keySerializer( - String keySerializer) { - doSetProperty("keySerializer", keySerializer); - return this; - } - /** - * The producer groups together any records that arrive in between - * request transmissions into a single batched request. Normally this - * occurs only under load when records arrive faster than they can be - * sent out. However in some circumstances the client may want to reduce - * the number of requests even under moderate load. This setting - * accomplishes this by adding a small amount of artificial - * delay&amp;mdash;that is, rather than immediately sending out a - * record the producer will wait for up to the given delay to allow - * other records to be sent so that the sends can be batched together. - * This can be thought of as analogous to Nagle's algorithm in TCP. This - * setting gives the upper bound on the delay for batching: once we get - * batch.size worth of records for a partition it will be sent - * immediately regardless of this setting, however if we have fewer than - * this many bytes accumulated for this partition we will 'linger' for - * the specified time waiting for more records to show up. This setting - * defaults to 0 (i.e. no delay). Setting linger.ms=5, for example, - * would have the effect of reducing the number of requests sent but - * would add up to 5ms of latency to records sent in the absence of - * load. - * - * The option is a: <code>long</code> type. - * - * Default: 0ms - * Group: producer - * - * @param lingerMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder lingerMs(long lingerMs) { - doSetProperty("lingerMs", lingerMs); - return this; - } - /** - * The producer groups together any records that arrive in between - * request transmissions into a single batched request. Normally this - * occurs only under load when records arrive faster than they can be - * sent out. However in some circumstances the client may want to reduce - * the number of requests even under moderate load. This setting - * accomplishes this by adding a small amount of artificial - * delay&amp;mdash;that is, rather than immediately sending out a - * record the producer will wait for up to the given delay to allow - * other records to be sent so that the sends can be batched together. - * This can be thought of as analogous to Nagle's algorithm in TCP. This - * setting gives the upper bound on the delay for batching: once we get - * batch.size worth of records for a partition it will be sent - * immediately regardless of this setting, however if we have fewer than - * this many bytes accumulated for this partition we will 'linger' for - * the specified time waiting for more records to show up. This setting - * defaults to 0 (i.e. no delay). Setting linger.ms=5, for example, - * would have the effect of reducing the number of requests sent but - * would add up to 5ms of latency to records sent in the absence of - * load. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 0ms - * Group: producer - * - * @param lingerMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder lingerMs(String lingerMs) { - doSetProperty("lingerMs", lingerMs); - return this; - } - /** - * The configuration controls how long the KafkaProducer's send(), - * partitionsFor(), initTransactions(), sendOffsetsToTransaction(), - * commitTransaction() and abortTransaction() methods will block. For - * send() this timeout bounds the total time waiting for both metadata - * fetch and buffer allocation (blocking in the user-supplied - * serializers or partitioner is not counted against this timeout). For - * partitionsFor() this timeout bounds the time spent waiting for - * metadata if it is unavailable. The transaction-related methods always - * block, but may timeout if the transaction coordinator could not be - * discovered or did not respond within the timeout. - * - * The option is a: <code>long</code> type. - * - * Default: 1m - * Group: producer - * - * @param maxBlockMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder maxBlockMs(long maxBlockMs) { - doSetProperty("maxBlockMs", maxBlockMs); - return this; - } - /** - * The configuration controls how long the KafkaProducer's send(), - * partitionsFor(), initTransactions(), sendOffsetsToTransaction(), - * commitTransaction() and abortTransaction() methods will block. For - * send() this timeout bounds the total time waiting for both metadata - * fetch and buffer allocation (blocking in the user-supplied - * serializers or partitioner is not counted against this timeout). For - * partitionsFor() this timeout bounds the time spent waiting for - * metadata if it is unavailable. The transaction-related methods always - * block, but may timeout if the transaction coordinator could not be - * discovered or did not respond within the timeout. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 1m - * Group: producer - * - * @param maxBlockMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder maxBlockMs(String maxBlockMs) { - doSetProperty("maxBlockMs", maxBlockMs); - return this; - } - /** - * The maximum number of unacknowledged requests the client will send on - * a single connection before blocking. Note that if this setting is set - * to be greater than 1 and there are failed sends, there is a risk of - * message re-ordering due to retries (i.e., if retries are enabled). - * - * The option is a: <code>int</code> type. - * - * Default: 5 - * Group: producer - * - * @param maxInFlightRequestsPerConnection the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder maxInFlightRequestsPerConnection( - int maxInFlightRequestsPerConnection) { - doSetProperty("maxInFlightRequestsPerConnection", maxInFlightRequestsPerConnection); - return this; - } - /** - * The maximum number of unacknowledged requests the client will send on - * a single connection before blocking. Note that if this setting is set - * to be greater than 1 and there are failed sends, there is a risk of - * message re-ordering due to retries (i.e., if retries are enabled). - * - * The option will be converted to a <code>int</code> type. - * - * Default: 5 - * Group: producer - * - * @param maxInFlightRequestsPerConnection the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder maxInFlightRequestsPerConnection( - String maxInFlightRequestsPerConnection) { - doSetProperty("maxInFlightRequestsPerConnection", maxInFlightRequestsPerConnection); - return this; - } - /** - * The maximum size of a request in bytes. This setting will limit the - * number of record batches the producer will send in a single request - * to avoid sending huge requests. This is also effectively a cap on the - * maximum uncompressed record batch size. Note that the server has its - * own cap on the record batch size (after compression if compression is - * enabled) which may be different from this. - * - * The option is a: <code>int</code> type. - * - * Default: 1048576 - * Group: producer - * - * @param maxRequestSize the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder maxRequestSize( - int maxRequestSize) { - doSetProperty("maxRequestSize", maxRequestSize); - return this; - } - /** - * The maximum size of a request in bytes. This setting will limit the - * number of record batches the producer will send in a single request - * to avoid sending huge requests. This is also effectively a cap on the - * maximum uncompressed record batch size. Note that the server has its - * own cap on the record batch size (after compression if compression is - * enabled) which may be different from this. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 1048576 - * Group: producer - * - * @param maxRequestSize the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder maxRequestSize( - String maxRequestSize) { - doSetProperty("maxRequestSize", maxRequestSize); - return this; - } - /** - * Controls how long the producer will cache metadata for a topic that's - * idle. If the elapsed time since a topic was last produced to exceeds - * the metadata idle duration, then the topic's metadata is forgotten - * and the next access to it will force a metadata fetch request. - * - * The option is a: <code>long</code> type. - * - * Default: 5m - * Group: producer - * - * @param metadataMaxIdleMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metadataMaxIdleMs( - long metadataMaxIdleMs) { - doSetProperty("metadataMaxIdleMs", metadataMaxIdleMs); - return this; - } - /** - * Controls how long the producer will cache metadata for a topic that's - * idle. If the elapsed time since a topic was last produced to exceeds - * the metadata idle duration, then the topic's metadata is forgotten - * and the next access to it will force a metadata fetch request. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 5m - * Group: producer - * - * @param metadataMaxIdleMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder metadataMaxIdleMs( - String metadataMaxIdleMs) { - doSetProperty("metadataMaxIdleMs", metadataMaxIdleMs); - return this; - } - /** - * Partitioner class that implements the - * org.apache.kafka.clients.producer.Partitioner interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: - * org.apache.kafka.clients.producer.internals.DefaultPartitioner - * Group: producer - * - * @param partitionerClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder partitionerClass( - String partitionerClass) { - doSetProperty("partitionerClass", partitionerClass); - return this; - } - /** - * Setting a value greater than zero will cause the client to resend any - * record whose send fails with a potentially transient error. Note that - * this retry is no different than if the client resent the record upon - * receiving the error. Allowing retries without setting - * max.in.flight.requests.per.connection to 1 will potentially change - * the ordering of records because if two batches are sent to a single - * partition, and the first fails and is retried but the second - * succeeds, then the records in the second batch may appear first. Note - * additionally that produce requests will be failed before the number - * of retries has been exhausted if the timeout configured by - * delivery.timeout.ms expires first before successful acknowledgement. - * Users should generally prefer to leave this config unset and instead - * use delivery.timeout.ms to control retry behavior. - * - * The option is a: <code>int</code> type. - * - * Default: 2147483647 - * Group: producer - * - * @param retries the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder retries(int retries) { - doSetProperty("retries", retries); - return this; - } - /** - * Setting a value greater than zero will cause the client to resend any - * record whose send fails with a potentially transient error. Note that - * this retry is no different than if the client resent the record upon - * receiving the error. Allowing retries without setting - * max.in.flight.requests.per.connection to 1 will potentially change - * the ordering of records because if two batches are sent to a single - * partition, and the first fails and is retried but the second - * succeeds, then the records in the second batch may appear first. Note - * additionally that produce requests will be failed before the number - * of retries has been exhausted if the timeout configured by - * delivery.timeout.ms expires first before successful acknowledgement. - * Users should generally prefer to leave this config unset and instead - * use delivery.timeout.ms to control retry behavior. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 2147483647 - * Group: producer - * - * @param retries the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder retries(String retries) { - doSetProperty("retries", retries); - return this; - } - /** - * The TransactionalId to use for transactional delivery. This enables - * reliability semantics which span multiple producer sessions since it - * allows the client to guarantee that transactions using the same - * TransactionalId have been completed prior to starting any new - * transactions. If no TransactionalId is provided, then the producer is - * limited to idempotent delivery. If a TransactionalId is configured, - * enable.idempotence is implied. By default the TransactionId is not - * configured, which means transactions cannot be used. Note that, by - * default, transactions require a cluster of at least three brokers - * which is the recommended setting for production; for development you - * can change this, by adjusting broker setting - * transaction.state.log.replication.factor. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: producer - * - * @param transactionalId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder transactionalId( - String transactionalId) { - doSetProperty("transactionalId", transactionalId); - return this; - } - /** - * The maximum amount of time in ms that the transaction coordinator - * will wait for a transaction status update from the producer before - * proactively aborting the ongoing transaction.If this value is larger - * than the transaction.max.timeout.ms setting in the broker, the - * request will fail with a InvalidTxnTimeoutException error. - * - * The option is a: <code>int</code> type. - * - * Default: 1m - * Group: producer - * - * @param transactionTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder transactionTimeoutMs( - int transactionTimeoutMs) { - doSetProperty("transactionTimeoutMs", transactionTimeoutMs); - return this; - } - /** - * The maximum amount of time in ms that the transaction coordinator - * will wait for a transaction status update from the producer before - * proactively aborting the ongoing transaction.If this value is larger - * than the transaction.max.timeout.ms setting in the broker, the - * request will fail with a InvalidTxnTimeoutException error. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 1m - * Group: producer - * - * @param transactionTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder transactionTimeoutMs( - String transactionTimeoutMs) { - doSetProperty("transactionTimeoutMs", transactionTimeoutMs); - return this; - } - /** - * Serializer class for value that implements the - * org.apache.kafka.common.serialization.Serializer interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: org.apache.kafka.common.serialization.StringSerializer - * Group: producer - * - * @param valueSerializer the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder valueSerializer( - String valueSerializer) { - doSetProperty("valueSerializer", valueSerializer); - return this; - } - /** - * The fully qualified name of a SASL client callback handler class that - * implements the AuthenticateCallbackHandler interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslClientCallbackHandlerClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslClientCallbackHandlerClass( - String saslClientCallbackHandlerClass) { - doSetProperty("saslClientCallbackHandlerClass", saslClientCallbackHandlerClass); - return this; - } - /** - * JAAS login context parameters for SASL connections in the format used - * by JAAS configuration files. JAAS configuration file format is - * described here. The format for the value is: loginModuleClass - * controlFlag (optionName=optionValue);. For brokers, the config must - * be prefixed with listener prefix and SASL mechanism name in - * lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=com.example.ScramLoginModule required;. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslJaasConfig the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslJaasConfig( - String saslJaasConfig) { - doSetProperty("saslJaasConfig", saslJaasConfig); - return this; - } - /** - * Kerberos kinit command path. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: /usr/bin/kinit - * Group: security - * - * @param saslKerberosKinitCmd the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslKerberosKinitCmd( - String saslKerberosKinitCmd) { - doSetProperty("saslKerberosKinitCmd", saslKerberosKinitCmd); - return this; - } - /** - * Login thread sleep time between refresh attempts. - * - * The option is a: <code>long</code> type. - * - * Default: 60000 - * Group: security - * - * @param saslKerberosMinTimeBeforeRelogin the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslKerberosMinTimeBeforeRelogin( - long saslKerberosMinTimeBeforeRelogin) { - doSetProperty("saslKerberosMinTimeBeforeRelogin", saslKerberosMinTimeBeforeRelogin); - return this; - } - /** - * Login thread sleep time between refresh attempts. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 60000 - * Group: security - * - * @param saslKerberosMinTimeBeforeRelogin the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslKerberosMinTimeBeforeRelogin( - String saslKerberosMinTimeBeforeRelogin) { - doSetProperty("saslKerberosMinTimeBeforeRelogin", saslKerberosMinTimeBeforeRelogin); - return this; - } - /** - * The Kerberos principal name that Kafka runs as. This can be defined - * either in Kafka's JAAS config or in Kafka's config. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslKerberosServiceName the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslKerberosServiceName( - String saslKerberosServiceName) { - doSetProperty("saslKerberosServiceName", saslKerberosServiceName); - return this; - } - /** - * Percentage of random jitter added to the renewal time. - * - * The option is a: <code>double</code> type. - * - * Default: 0.05 - * Group: security - * - * @param saslKerberosTicketRenewJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslKerberosTicketRenewJitter( - double saslKerberosTicketRenewJitter) { - doSetProperty("saslKerberosTicketRenewJitter", saslKerberosTicketRenewJitter); - return this; - } - /** - * Percentage of random jitter added to the renewal time. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.05 - * Group: security - * - * @param saslKerberosTicketRenewJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslKerberosTicketRenewJitter( - String saslKerberosTicketRenewJitter) { - doSetProperty("saslKerberosTicketRenewJitter", saslKerberosTicketRenewJitter); - return this; - } - /** - * Login thread will sleep until the specified window factor of time - * from last refresh to ticket's expiry has been reached, at which time - * it will try to renew the ticket. - * - * The option is a: <code>double</code> type. - * - * Default: 0.8 - * Group: security - * - * @param saslKerberosTicketRenewWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslKerberosTicketRenewWindowFactor( - double saslKerberosTicketRenewWindowFactor) { - doSetProperty("saslKerberosTicketRenewWindowFactor", saslKerberosTicketRenewWindowFactor); - return this; - } - /** - * Login thread will sleep until the specified window factor of time - * from last refresh to ticket's expiry has been reached, at which time - * it will try to renew the ticket. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.8 - * Group: security - * - * @param saslKerberosTicketRenewWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslKerberosTicketRenewWindowFactor( - String saslKerberosTicketRenewWindowFactor) { - doSetProperty("saslKerberosTicketRenewWindowFactor", saslKerberosTicketRenewWindowFactor); - return this; - } - /** - * The fully qualified name of a SASL login callback handler class that - * implements the AuthenticateCallbackHandler interface. For brokers, - * login callback handler config must be prefixed with listener prefix - * and SASL mechanism name in lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.login.callback.handler.class=com.example.CustomScramLoginCallbackHandler. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslLoginCallbackHandlerClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginCallbackHandlerClass( - String saslLoginCallbackHandlerClass) { - doSetProperty("saslLoginCallbackHandlerClass", saslLoginCallbackHandlerClass); - return this; - } - /** - * The fully qualified name of a class that implements the Login - * interface. For brokers, login config must be prefixed with listener - * prefix and SASL mechanism name in lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.login.class=com.example.CustomScramLogin. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslLoginClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginClass( - String saslLoginClass) { - doSetProperty("saslLoginClass", saslLoginClass); - return this; - } - /** - * The amount of buffer time before credential expiration to maintain - * when refreshing a credential, in seconds. If a refresh would - * otherwise occur closer to expiration than the number of buffer - * seconds then the refresh will be moved up to maintain as much of the - * buffer time as possible. Legal values are between 0 and 3600 (1 - * hour); a default value of 300 (5 minutes) is used if no value is - * specified. This value and sasl.login.refresh.min.period.seconds are - * both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>short</code> type. - * - * Default: 300 - * Group: security - * - * @param saslLoginRefreshBufferSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginRefreshBufferSeconds( - short saslLoginRefreshBufferSeconds) { - doSetProperty("saslLoginRefreshBufferSeconds", saslLoginRefreshBufferSeconds); - return this; - } - /** - * The amount of buffer time before credential expiration to maintain - * when refreshing a credential, in seconds. If a refresh would - * otherwise occur closer to expiration than the number of buffer - * seconds then the refresh will be moved up to maintain as much of the - * buffer time as possible. Legal values are between 0 and 3600 (1 - * hour); a default value of 300 (5 minutes) is used if no value is - * specified. This value and sasl.login.refresh.min.period.seconds are - * both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>short</code> - * type. - * - * Default: 300 - * Group: security - * - * @param saslLoginRefreshBufferSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginRefreshBufferSeconds( - String saslLoginRefreshBufferSeconds) { - doSetProperty("saslLoginRefreshBufferSeconds", saslLoginRefreshBufferSeconds); - return this; - } - /** - * The desired minimum time for the login refresh thread to wait before - * refreshing a credential, in seconds. Legal values are between 0 and - * 900 (15 minutes); a default value of 60 (1 minute) is used if no - * value is specified. This value and sasl.login.refresh.buffer.seconds - * are both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>short</code> type. - * - * Default: 60 - * Group: security - * - * @param saslLoginRefreshMinPeriodSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginRefreshMinPeriodSeconds( - short saslLoginRefreshMinPeriodSeconds) { - doSetProperty("saslLoginRefreshMinPeriodSeconds", saslLoginRefreshMinPeriodSeconds); - return this; - } - /** - * The desired minimum time for the login refresh thread to wait before - * refreshing a credential, in seconds. Legal values are between 0 and - * 900 (15 minutes); a default value of 60 (1 minute) is used if no - * value is specified. This value and sasl.login.refresh.buffer.seconds - * are both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>short</code> - * type. - * - * Default: 60 - * Group: security - * - * @param saslLoginRefreshMinPeriodSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginRefreshMinPeriodSeconds( - String saslLoginRefreshMinPeriodSeconds) { - doSetProperty("saslLoginRefreshMinPeriodSeconds", saslLoginRefreshMinPeriodSeconds); - return this; - } - /** - * Login refresh thread will sleep until the specified window factor - * relative to the credential's lifetime has been reached, at which time - * it will try to refresh the credential. Legal values are between 0.5 - * (50%) and 1.0 (100%) inclusive; a default value of 0.8 (80%) is used - * if no value is specified. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>double</code> type. - * - * Default: 0.8 - * Group: security - * - * @param saslLoginRefreshWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginRefreshWindowFactor( - double saslLoginRefreshWindowFactor) { - doSetProperty("saslLoginRefreshWindowFactor", saslLoginRefreshWindowFactor); - return this; - } - /** - * Login refresh thread will sleep until the specified window factor - * relative to the credential's lifetime has been reached, at which time - * it will try to refresh the credential. Legal values are between 0.5 - * (50%) and 1.0 (100%) inclusive; a default value of 0.8 (80%) is used - * if no value is specified. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.8 - * Group: security - * - * @param saslLoginRefreshWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginRefreshWindowFactor( - String saslLoginRefreshWindowFactor) { - doSetProperty("saslLoginRefreshWindowFactor", saslLoginRefreshWindowFactor); - return this; - } - /** - * The maximum amount of random jitter relative to the credential's - * lifetime that is added to the login refresh thread's sleep time. - * Legal values are between 0 and 0.25 (25%) inclusive; a default value - * of 0.05 (5%) is used if no value is specified. Currently applies only - * to OAUTHBEARER. - * - * The option is a: <code>double</code> type. - * - * Default: 0.05 - * Group: security - * - * @param saslLoginRefreshWindowJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginRefreshWindowJitter( - double saslLoginRefreshWindowJitter) { - doSetProperty("saslLoginRefreshWindowJitter", saslLoginRefreshWindowJitter); - return this; - } - /** - * The maximum amount of random jitter relative to the credential's - * lifetime that is added to the login refresh thread's sleep time. - * Legal values are between 0 and 0.25 (25%) inclusive; a default value - * of 0.05 (5%) is used if no value is specified. Currently applies only - * to OAUTHBEARER. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.05 - * Group: security - * - * @param saslLoginRefreshWindowJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslLoginRefreshWindowJitter( - String saslLoginRefreshWindowJitter) { - doSetProperty("saslLoginRefreshWindowJitter", saslLoginRefreshWindowJitter); - return this; - } - /** - * SASL mechanism used for client connections. This may be any mechanism - * for which a security provider is available. GSSAPI is the default - * mechanism. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: GSSAPI - * Group: security - * - * @param saslMechanism the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder saslMechanism( - String saslMechanism) { - doSetProperty("saslMechanism", saslMechanism); - return this; - } - /** - * Protocol used to communicate with brokers. Valid values are: - * PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: PLAINTEXT - * Group: security - * - * @param securityProtocol the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder securityProtocol( - String securityProtocol) { - doSetProperty("securityProtocol", securityProtocol); - return this; - } - /** - * A list of configurable creator classes each returning a provider - * implementing security algorithms. These classes should implement the - * org.apache.kafka.common.security.auth.SecurityProviderCreator - * interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param securityProviders the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder securityProviders( - String securityProviders) { - doSetProperty("securityProviders", securityProviders); - return this; - } - /** - * A list of cipher suites. This is a named combination of - * authentication, encryption, MAC and key exchange algorithm used to - * negotiate the security settings for a network connection using TLS or - * SSL network protocol. By default all the available cipher suites are - * supported. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslCipherSuites the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslCipherSuites( - String sslCipherSuites) { - doSetProperty("sslCipherSuites", sslCipherSuites); - return this; - } - /** - * The list of protocols enabled for SSL connections. The default is - * 'TLSv1.2,TLSv1.3' when running with Java 11 or newer, 'TLSv1.2' - * otherwise. With the default value for Java 11, clients and servers - * will prefer TLSv1.3 if both support it and fallback to TLSv1.2 - * otherwise (assuming both support at least TLSv1.2). This default - * should be fine for most cases. Also see the config documentation for - * ssl.protocol. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: TLSv1.2,TLSv1.3 - * Group: security - * - * @param sslEnabledProtocols the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslEnabledProtocols( - String sslEnabledProtocols) { - doSetProperty("sslEnabledProtocols", sslEnabledProtocols); - return this; - } - /** - * The endpoint identification algorithm to validate server hostname - * using server certificate. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: https - * Group: security - * - * @param sslEndpointIdentificationAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslEndpointIdentificationAlgorithm( - String sslEndpointIdentificationAlgorithm) { - doSetProperty("sslEndpointIdentificationAlgorithm", sslEndpointIdentificationAlgorithm); - return this; - } - /** - * The class of type - * org.apache.kafka.common.security.auth.SslEngineFactory to provide - * SSLEngine objects. Default value is - * org.apache.kafka.common.security.ssl.DefaultSslEngineFactory. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslEngineFactoryClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslEngineFactoryClass( - String sslEngineFactoryClass) { - doSetProperty("sslEngineFactoryClass", sslEngineFactoryClass); - return this; - } - /** - * The algorithm used by key manager factory for SSL connections. - * Default value is the key manager factory algorithm configured for the - * Java Virtual Machine. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: SunX509 - * Group: security - * - * @param sslKeymanagerAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslKeymanagerAlgorithm( - String sslKeymanagerAlgorithm) { - doSetProperty("sslKeymanagerAlgorithm", sslKeymanagerAlgorithm); - return this; - } - /** - * The password of the private key in the key store file orthe PEM key - * specified in ssl.keystore.key'. This is required for clients only if - * two-way authentication is configured. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeyPassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslKeyPassword( - String sslKeyPassword) { - doSetProperty("sslKeyPassword", sslKeyPassword); - return this; - } - /** - * Certificate chain in the format specified by 'ssl.keystore.type'. - * Default SSL engine factory supports only PEM format with a list of - * X.509 certificates. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreCertificateChain the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslKeystoreCertificateChain( - String sslKeystoreCertificateChain) { - doSetProperty("sslKeystoreCertificateChain", sslKeystoreCertificateChain); - return this; - } - /** - * Private key in the format specified by 'ssl.keystore.type'. Default - * SSL engine factory supports only PEM format with PKCS#8 keys. If the - * key is encrypted, key password must be specified using - * 'ssl.key.password'. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreKey the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslKeystoreKey( - String sslKeystoreKey) { - doSetProperty("sslKeystoreKey", sslKeystoreKey); - return this; - } - /** - * The location of the key store file. This is optional for client and - * can be used for two-way authentication for client. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreLocation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslKeystoreLocation( - String sslKeystoreLocation) { - doSetProperty("sslKeystoreLocation", sslKeystoreLocation); - return this; - } - /** - * The store password for the key store file. This is optional for - * client and only needed if 'ssl.keystore.location' is configured. Key - * store password is not supported for PEM format. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystorePassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslKeystorePassword( - String sslKeystorePassword) { - doSetProperty("sslKeystorePassword", sslKeystorePassword); - return this; - } - /** - * The file format of the key store file. This is optional for client. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: JKS - * Group: security - * - * @param sslKeystoreType the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslKeystoreType( - String sslKeystoreType) { - doSetProperty("sslKeystoreType", sslKeystoreType); - return this; - } - /** - * The SSL protocol used to generate the SSLContext. The default is - * 'TLSv1.3' when running with Java 11 or newer, 'TLSv1.2' otherwise. - * This value should be fine for most use cases. Allowed values in - * recent JVMs are 'TLSv1.2' and 'TLSv1.3'. 'TLS', 'TLSv1.1', 'SSL', - * 'SSLv2' and 'SSLv3' may be supported in older JVMs, but their usage - * is discouraged due to known security vulnerabilities. With the - * default value for this config and 'ssl.enabled.protocols', clients - * will downgrade to 'TLSv1.2' if the server does not support 'TLSv1.3'. - * If this config is set to 'TLSv1.2', clients will not use 'TLSv1.3' - * even if it is one of the values in ssl.enabled.protocols and the - * server only supports 'TLSv1.3'. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: TLSv1.2 - * Group: security - * - * @param sslProtocol the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslProtocol(String sslProtocol) { - doSetProperty("sslProtocol", sslProtocol); - return this; - } - /** - * The name of the security provider used for SSL connections. Default - * value is the default security provider of the JVM. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslProvider the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslProvider(String sslProvider) { - doSetProperty("sslProvider", sslProvider); - return this; - } - /** - * The SecureRandom PRNG implementation to use for SSL cryptography - * operations. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslSecureRandomImplementation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslSecureRandomImplementation( - String sslSecureRandomImplementation) { - doSetProperty("sslSecureRandomImplementation", sslSecureRandomImplementation); - return this; - } - /** - * The algorithm used by trust manager factory for SSL connections. - * Default value is the trust manager factory algorithm configured for - * the Java Virtual Machine. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: PKIX - * Group: security - * - * @param sslTrustmanagerAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslTrustmanagerAlgorithm( - String sslTrustmanagerAlgorithm) { - doSetProperty("sslTrustmanagerAlgorithm", sslTrustmanagerAlgorithm); - return this; - } - /** - * Trusted certificates in the format specified by - * 'ssl.truststore.type'. Default SSL engine factory supports only PEM - * format with X.509 certificates. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststoreCertificates the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslTruststoreCertificates( - String sslTruststoreCertificates) { - doSetProperty("sslTruststoreCertificates", sslTruststoreCertificates); - return this; - } - /** - * The location of the trust store file. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststoreLocation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslTruststoreLocation( - String sslTruststoreLocation) { - doSetProperty("sslTruststoreLocation", sslTruststoreLocation); - return this; - } - /** - * The password for the trust store file. If a password is not set, - * trust store file configured will still be used, but integrity - * checking is disabled. Trust store password is not supported for PEM - * format. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststorePassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslTruststorePassword( - String sslTruststorePassword) { - doSetProperty("sslTruststorePassword", sslTruststorePassword); - return this; - } - /** - * The file format of the trust store file. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: JKS - * Group: security - * - * @param sslTruststoreType the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointProducerBuilder sslTruststoreType( - String sslTruststoreType) { - doSetProperty("sslTruststoreType", sslTruststoreType); - return this; - } - } - - /** - * Advanced builder for endpoint producers for the Vert.x Kafka component. - */ - public interface AdvancedVertxKafkaEndpointProducerBuilder - extends - EndpointProducerBuilder { - default VertxKafkaEndpointProducerBuilder basic() { - return (VertxKafkaEndpointProducerBuilder) this; - } - /** - * Whether the producer should be started lazy (on the first message). - * By starting lazy you can use this to allow CamelContext and routes to - * startup in situations where a producer may otherwise fail during - * starting and cause the route to fail being started. By deferring this - * startup to be lazy then the startup failure can be handled during - * routing messages via Camel's routing error handlers. Beware that when - * the first message is processed then creating and starting the - * producer may take a little time and prolong the total processing time - * of the processing. - * - * The option is a: <code>boolean</code> type. - * - * Default: false - * Group: producer (advanced) - * - * @param lazyStartProducer the value to set - * @return the dsl builder - */ - default AdvancedVertxKafkaEndpointProducerBuilder lazyStartProducer( - boolean lazyStartProducer) { - doSetProperty("lazyStartProducer", lazyStartProducer); - return this; - } - /** - * Whether the producer should be started lazy (on the first message). - * By starting lazy you can use this to allow CamelContext and routes to - * startup in situations where a producer may otherwise fail during - * starting and cause the route to fail being started. By deferring this - * startup to be lazy then the startup failure can be handled during - * routing messages via Camel's routing error handlers. Beware that when - * the first message is processed then creating and starting the - * producer may take a little time and prolong the total processing time - * of the processing. - * - * The option will be converted to a <code>boolean</code> - * type. - * - * Default: false - * Group: producer (advanced) - * - * @param lazyStartProducer the value to set - * @return the dsl builder - */ - default AdvancedVertxKafkaEndpointProducerBuilder lazyStartProducer( - String lazyStartProducer) { - doSetProperty("lazyStartProducer", lazyStartProducer); - return this; - } - } - - /** - * Builder for endpoint for the Vert.x Kafka component. - */ - public interface VertxKafkaEndpointBuilder - extends - VertxKafkaEndpointConsumerBuilder, - VertxKafkaEndpointProducerBuilder { - default AdvancedVertxKafkaEndpointBuilder advanced() { - return (AdvancedVertxKafkaEndpointBuilder) this; - } - /** - * Sets additional properties for either kafka consumer or kafka - * producer in case they can't be set directly on the camel - * configurations (e.g: new Kafka properties that are not reflected yet - * in Camel configurations), the properties have to be prefixed with - * additionalProperties.. E.g: - * additionalProperties.transactional.id=12345&amp;additionalProperties.schema.registry.url=http://localhost:8811/avro. - * - * The option is a: <code>java.util.Map&lt;java.lang.String, - * java.lang.Object&gt;</code> type. - * The option is multivalued, and you can use the - * additionalProperties(String, Object) method to add a value (call the - * method multiple times to set more values). - * - * Group: common - * - * @param key the option key - * @param value the option value - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder additionalProperties( - String key, - Object value) { - doSetMultiValueProperty("additionalProperties", "additionalProperties." + key, value); - return this; - } - /** - * Sets additional properties for either kafka consumer or kafka - * producer in case they can't be set directly on the camel - * configurations (e.g: new Kafka properties that are not reflected yet - * in Camel configurations), the properties have to be prefixed with - * additionalProperties.. E.g: - * additionalProperties.transactional.id=12345&amp;additionalProperties.schema.registry.url=http://localhost:8811/avro. - * - * The option is a: <code>java.util.Map&lt;java.lang.String, - * java.lang.Object&gt;</code> type. - * The option is multivalued, and you can use the - * additionalProperties(String, Object) method to add a value (call the - * method multiple times to set more values). - * - * Group: common - * - * @param values the values - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder additionalProperties(Map values) { - doSetMultiValueProperties("additionalProperties", "additionalProperties.", values); - return this; - } - /** - * A list of host/port pairs to use for establishing the initial - * connection to the Kafka cluster. The client will make use of all - * servers irrespective of which servers are specified here for - * bootstrapping&amp;mdash;this list only impacts the initial hosts - * used to discover the full set of servers. This list should be in the - * form host1:port1,host2:port2,.... Since these servers are just used - * for the initial connection to discover the full cluster membership - * (which may change dynamically), this list need not contain the full - * set of servers (you may want more than one, though, in case a server - * is down). - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param bootstrapServers the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder bootstrapServers( - String bootstrapServers) { - doSetProperty("bootstrapServers", bootstrapServers); - return this; - } - /** - * Controls how the client uses DNS lookups. If set to use_all_dns_ips, - * connect to each returned IP address in sequence until a successful - * connection is established. After a disconnection, the next IP is - * used. Once all IPs have been used once, the client resolves the IP(s) - * from the hostname again (both the JVM and the OS cache DNS name - * lookups, however). If set to - * resolve_canonical_bootstrap_servers_only, resolve each bootstrap - * address into a list of canonical names. After the bootstrap phase, - * this behaves the same as use_all_dns_ips. If set to default - * (deprecated), attempt to connect to the first IP address returned by - * the lookup, even if the lookup returns multiple IP addresses. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: use_all_dns_ips - * Group: common - * - * @param clientDnsLookup the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder clientDnsLookup(String clientDnsLookup) { - doSetProperty("clientDnsLookup", clientDnsLookup); - return this; - } - /** - * An id string to pass to the server when making requests. The purpose - * of this is to be able to track the source of requests beyond just - * ip/port by allowing a logical application name to be included in - * server-side request logging. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param clientId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder clientId(String clientId) { - doSetProperty("clientId", clientId); - return this; - } - /** - * Close idle connections after the number of milliseconds specified by - * this config. - * - * The option is a: <code>long</code> type. - * - * Default: 9m - * Group: common - * - * @param connectionsMaxIdleMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder connectionsMaxIdleMs( - long connectionsMaxIdleMs) { - doSetProperty("connectionsMaxIdleMs", connectionsMaxIdleMs); - return this; - } - /** - * Close idle connections after the number of milliseconds specified by - * this config. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 9m - * Group: common - * - * @param connectionsMaxIdleMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder connectionsMaxIdleMs( - String connectionsMaxIdleMs) { - doSetProperty("connectionsMaxIdleMs", connectionsMaxIdleMs); - return this; - } - /** - * To use a custom HeaderFilterStrategy to filter header to and from - * Camel message. - * - * The option is a: - * <code>org.apache.camel.spi.HeaderFilterStrategy</code> - * type. - * - * Group: common - * - * @param headerFilterStrategy the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder headerFilterStrategy( - org.apache.camel.spi.HeaderFilterStrategy headerFilterStrategy) { - doSetProperty("headerFilterStrategy", headerFilterStrategy); - return this; - } - /** - * To use a custom HeaderFilterStrategy to filter header to and from - * Camel message. - * - * The option will be converted to a - * <code>org.apache.camel.spi.HeaderFilterStrategy</code> - * type. - * - * Group: common - * - * @param headerFilterStrategy the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder headerFilterStrategy( - String headerFilterStrategy) { - doSetProperty("headerFilterStrategy", headerFilterStrategy); - return this; - } - /** - * A list of classes to use as interceptors. Implementing the - * org.apache.kafka.clients.producer.ProducerInterceptor interface - * allows you to intercept (and possibly mutate) the records received by - * the producer before they are published to the Kafka cluster. By - * default, there are no interceptors. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param interceptorClasses the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder interceptorClasses( - String interceptorClasses) { - doSetProperty("interceptorClasses", interceptorClasses); - return this; - } - /** - * The period of time in milliseconds after which we force a refresh of - * metadata even if we haven't seen any partition leadership changes to - * proactively discover any new brokers or partitions. - * - * The option is a: <code>long</code> type. - * - * Default: 5m - * Group: common - * - * @param metadataMaxAgeMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder metadataMaxAgeMs(long metadataMaxAgeMs) { - doSetProperty("metadataMaxAgeMs", metadataMaxAgeMs); - return this; - } - /** - * The period of time in milliseconds after which we force a refresh of - * metadata even if we haven't seen any partition leadership changes to - * proactively discover any new brokers or partitions. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 5m - * Group: common - * - * @param metadataMaxAgeMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder metadataMaxAgeMs( - String metadataMaxAgeMs) { - doSetProperty("metadataMaxAgeMs", metadataMaxAgeMs); - return this; - } - /** - * A list of classes to use as metrics reporters. Implementing the - * org.apache.kafka.common.metrics.MetricsReporter interface allows - * plugging in classes that will be notified of new metric creation. The - * JmxReporter is always included to register JMX statistics. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: common - * - * @param metricReporters the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder metricReporters(String metricReporters) { - doSetProperty("metricReporters", metricReporters); - return this; - } - /** - * The number of samples maintained to compute metrics. - * - * The option is a: <code>int</code> type. - * - * Default: 2 - * Group: common - * - * @param metricsNumSamples the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder metricsNumSamples( - int metricsNumSamples) { - doSetProperty("metricsNumSamples", metricsNumSamples); - return this; - } - /** - * The number of samples maintained to compute metrics. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 2 - * Group: common - * - * @param metricsNumSamples the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder metricsNumSamples( - String metricsNumSamples) { - doSetProperty("metricsNumSamples", metricsNumSamples); - return this; - } - /** - * The highest recording level for metrics. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: INFO - * Group: common - * - * @param metricsRecordingLevel the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder metricsRecordingLevel( - String metricsRecordingLevel) { - doSetProperty("metricsRecordingLevel", metricsRecordingLevel); - return this; - } - /** - * The window of time a metrics sample is computed over. - * - * The option is a: <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param metricsSampleWindowMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder metricsSampleWindowMs( - long metricsSampleWindowMs) { - doSetProperty("metricsSampleWindowMs", metricsSampleWindowMs); - return this; - } - /** - * The window of time a metrics sample is computed over. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param metricsSampleWindowMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder metricsSampleWindowMs( - String metricsSampleWindowMs) { - doSetProperty("metricsSampleWindowMs", metricsSampleWindowMs); - return this; - } - /** - * The partition to which the record will be sent (or null if no - * partition was specified) or read from a particular partition if set. - * Header VertxKafkaConstants#PARTITION_ID If configured, it will take - * precedence over this config. - * - * The option is a: <code>java.lang.Integer</code> type. - * - * Group: common - * - * @param partitionId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder partitionId(Integer partitionId) { - doSetProperty("partitionId", partitionId); - return this; - } - /** - * The partition to which the record will be sent (or null if no - * partition was specified) or read from a particular partition if set. - * Header VertxKafkaConstants#PARTITION_ID If configured, it will take - * precedence over this config. - * - * The option will be converted to a - * <code>java.lang.Integer</code> type. - * - * Group: common - * - * @param partitionId the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder partitionId(String partitionId) { - doSetProperty("partitionId", partitionId); - return this; - } - /** - * The size of the TCP receive buffer (SO_RCVBUF) to use when reading - * data. If the value is -1, the OS default will be used. - * - * The option is a: <code>int</code> type. - * - * Default: 32768 - * Group: common - * - * @param receiveBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder receiveBufferBytes( - int receiveBufferBytes) { - doSetProperty("receiveBufferBytes", receiveBufferBytes); - return this; - } - /** - * The size of the TCP receive buffer (SO_RCVBUF) to use when reading - * data. If the value is -1, the OS default will be used. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 32768 - * Group: common - * - * @param receiveBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder receiveBufferBytes( - String receiveBufferBytes) { - doSetProperty("receiveBufferBytes", receiveBufferBytes); - return this; - } - /** - * The maximum amount of time in milliseconds to wait when reconnecting - * to a broker that has repeatedly failed to connect. If provided, the - * backoff per host will increase exponentially for each consecutive - * connection failure, up to this maximum. After calculating the backoff - * increase, 20% random jitter is added to avoid connection storms. - * - * The option is a: <code>long</code> type. - * - * Default: 1s - * Group: common - * - * @param reconnectBackoffMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder reconnectBackoffMaxMs( - long reconnectBackoffMaxMs) { - doSetProperty("reconnectBackoffMaxMs", reconnectBackoffMaxMs); - return this; - } - /** - * The maximum amount of time in milliseconds to wait when reconnecting - * to a broker that has repeatedly failed to connect. If provided, the - * backoff per host will increase exponentially for each consecutive - * connection failure, up to this maximum. After calculating the backoff - * increase, 20% random jitter is added to avoid connection storms. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 1s - * Group: common - * - * @param reconnectBackoffMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder reconnectBackoffMaxMs( - String reconnectBackoffMaxMs) { - doSetProperty("reconnectBackoffMaxMs", reconnectBackoffMaxMs); - return this; - } - /** - * The base amount of time to wait before attempting to reconnect to a - * given host. This avoids repeatedly connecting to a host in a tight - * loop. This backoff applies to all connection attempts by the client - * to a broker. - * - * The option is a: <code>long</code> type. - * - * Default: 50ms - * Group: common - * - * @param reconnectBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder reconnectBackoffMs( - long reconnectBackoffMs) { - doSetProperty("reconnectBackoffMs", reconnectBackoffMs); - return this; - } - /** - * The base amount of time to wait before attempting to reconnect to a - * given host. This avoids repeatedly connecting to a host in a tight - * loop. This backoff applies to all connection attempts by the client - * to a broker. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 50ms - * Group: common - * - * @param reconnectBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder reconnectBackoffMs( - String reconnectBackoffMs) { - doSetProperty("reconnectBackoffMs", reconnectBackoffMs); - return this; - } - /** - * The configuration controls the maximum amount of time the client will - * wait for the response of a request. If the response is not received - * before the timeout elapses the client will resend the request if - * necessary or fail the request if retries are exhausted. This should - * be larger than replica.lag.time.max.ms (a broker configuration) to - * reduce the possibility of message duplication due to unnecessary - * producer retries. - * - * The option is a: <code>int</code> type. - * - * Default: 30s - * Group: common - * - * @param requestTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder requestTimeoutMs(int requestTimeoutMs) { - doSetProperty("requestTimeoutMs", requestTimeoutMs); - return this; - } - /** - * The configuration controls the maximum amount of time the client will - * wait for the response of a request. If the response is not received - * before the timeout elapses the client will resend the request if - * necessary or fail the request if retries are exhausted. This should - * be larger than replica.lag.time.max.ms (a broker configuration) to - * reduce the possibility of message duplication due to unnecessary - * producer retries. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 30s - * Group: common - * - * @param requestTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder requestTimeoutMs( - String requestTimeoutMs) { - doSetProperty("requestTimeoutMs", requestTimeoutMs); - return this; - } - /** - * The amount of time to wait before attempting to retry a failed - * request to a given topic partition. This avoids repeatedly sending - * requests in a tight loop under some failure scenarios. - * - * The option is a: <code>long</code> type. - * - * Default: 100ms - * Group: common - * - * @param retryBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder retryBackoffMs(long retryBackoffMs) { - doSetProperty("retryBackoffMs", retryBackoffMs); - return this; - } - /** - * The amount of time to wait before attempting to retry a failed - * request to a given topic partition. This avoids repeatedly sending - * requests in a tight loop under some failure scenarios. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 100ms - * Group: common - * - * @param retryBackoffMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder retryBackoffMs(String retryBackoffMs) { - doSetProperty("retryBackoffMs", retryBackoffMs); - return this; - } - /** - * The size of the TCP send buffer (SO_SNDBUF) to use when sending data. - * If the value is -1, the OS default will be used. - * - * The option is a: <code>int</code> type. - * - * Default: 131072 - * Group: common - * - * @param sendBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sendBufferBytes(int sendBufferBytes) { - doSetProperty("sendBufferBytes", sendBufferBytes); - return this; - } - /** - * The size of the TCP send buffer (SO_SNDBUF) to use when sending data. - * If the value is -1, the OS default will be used. - * - * The option will be converted to a <code>int</code> type. - * - * Default: 131072 - * Group: common - * - * @param sendBufferBytes the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sendBufferBytes(String sendBufferBytes) { - doSetProperty("sendBufferBytes", sendBufferBytes); - return this; - } - /** - * The maximum amount of time the client will wait for the socket - * connection to be established. The connection setup timeout will - * increase exponentially for each consecutive connection failure up to - * this maximum. To avoid connection storms, a randomization factor of - * 0.2 will be applied to the timeout resulting in a random range - * between 20% below and 20% above the computed value. - * - * The option is a: <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param socketConnectionSetupTimeoutMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder socketConnectionSetupTimeoutMaxMs( - long socketConnectionSetupTimeoutMaxMs) { - doSetProperty("socketConnectionSetupTimeoutMaxMs", socketConnectionSetupTimeoutMaxMs); - return this; - } - /** - * The maximum amount of time the client will wait for the socket - * connection to be established. The connection setup timeout will - * increase exponentially for each consecutive connection failure up to - * this maximum. To avoid connection storms, a randomization factor of - * 0.2 will be applied to the timeout resulting in a random range - * between 20% below and 20% above the computed value. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 30s - * Group: common - * - * @param socketConnectionSetupTimeoutMaxMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder socketConnectionSetupTimeoutMaxMs( - String socketConnectionSetupTimeoutMaxMs) { - doSetProperty("socketConnectionSetupTimeoutMaxMs", socketConnectionSetupTimeoutMaxMs); - return this; - } - /** - * The amount of time the client will wait for the socket connection to - * be established. If the connection is not built before the timeout - * elapses, clients will close the socket channel. - * - * The option is a: <code>long</code> type. - * - * Default: 10s - * Group: common - * - * @param socketConnectionSetupTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder socketConnectionSetupTimeoutMs( - long socketConnectionSetupTimeoutMs) { - doSetProperty("socketConnectionSetupTimeoutMs", socketConnectionSetupTimeoutMs); - return this; - } - /** - * The amount of time the client will wait for the socket connection to - * be established. If the connection is not built before the timeout - * elapses, clients will close the socket channel. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 10s - * Group: common - * - * @param socketConnectionSetupTimeoutMs the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder socketConnectionSetupTimeoutMs( - String socketConnectionSetupTimeoutMs) { - doSetProperty("socketConnectionSetupTimeoutMs", socketConnectionSetupTimeoutMs); - return this; - } - /** - * The fully qualified name of a SASL client callback handler class that - * implements the AuthenticateCallbackHandler interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslClientCallbackHandlerClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslClientCallbackHandlerClass( - String saslClientCallbackHandlerClass) { - doSetProperty("saslClientCallbackHandlerClass", saslClientCallbackHandlerClass); - return this; - } - /** - * JAAS login context parameters for SASL connections in the format used - * by JAAS configuration files. JAAS configuration file format is - * described here. The format for the value is: loginModuleClass - * controlFlag (optionName=optionValue);. For brokers, the config must - * be prefixed with listener prefix and SASL mechanism name in - * lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=com.example.ScramLoginModule required;. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslJaasConfig the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslJaasConfig(String saslJaasConfig) { - doSetProperty("saslJaasConfig", saslJaasConfig); - return this; - } - /** - * Kerberos kinit command path. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: /usr/bin/kinit - * Group: security - * - * @param saslKerberosKinitCmd the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslKerberosKinitCmd( - String saslKerberosKinitCmd) { - doSetProperty("saslKerberosKinitCmd", saslKerberosKinitCmd); - return this; - } - /** - * Login thread sleep time between refresh attempts. - * - * The option is a: <code>long</code> type. - * - * Default: 60000 - * Group: security - * - * @param saslKerberosMinTimeBeforeRelogin the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslKerberosMinTimeBeforeRelogin( - long saslKerberosMinTimeBeforeRelogin) { - doSetProperty("saslKerberosMinTimeBeforeRelogin", saslKerberosMinTimeBeforeRelogin); - return this; - } - /** - * Login thread sleep time between refresh attempts. - * - * The option will be converted to a <code>long</code> type. - * - * Default: 60000 - * Group: security - * - * @param saslKerberosMinTimeBeforeRelogin the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslKerberosMinTimeBeforeRelogin( - String saslKerberosMinTimeBeforeRelogin) { - doSetProperty("saslKerberosMinTimeBeforeRelogin", saslKerberosMinTimeBeforeRelogin); - return this; - } - /** - * The Kerberos principal name that Kafka runs as. This can be defined - * either in Kafka's JAAS config or in Kafka's config. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslKerberosServiceName the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslKerberosServiceName( - String saslKerberosServiceName) { - doSetProperty("saslKerberosServiceName", saslKerberosServiceName); - return this; - } - /** - * Percentage of random jitter added to the renewal time. - * - * The option is a: <code>double</code> type. - * - * Default: 0.05 - * Group: security - * - * @param saslKerberosTicketRenewJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslKerberosTicketRenewJitter( - double saslKerberosTicketRenewJitter) { - doSetProperty("saslKerberosTicketRenewJitter", saslKerberosTicketRenewJitter); - return this; - } - /** - * Percentage of random jitter added to the renewal time. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.05 - * Group: security - * - * @param saslKerberosTicketRenewJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslKerberosTicketRenewJitter( - String saslKerberosTicketRenewJitter) { - doSetProperty("saslKerberosTicketRenewJitter", saslKerberosTicketRenewJitter); - return this; - } - /** - * Login thread will sleep until the specified window factor of time - * from last refresh to ticket's expiry has been reached, at which time - * it will try to renew the ticket. - * - * The option is a: <code>double</code> type. - * - * Default: 0.8 - * Group: security - * - * @param saslKerberosTicketRenewWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslKerberosTicketRenewWindowFactor( - double saslKerberosTicketRenewWindowFactor) { - doSetProperty("saslKerberosTicketRenewWindowFactor", saslKerberosTicketRenewWindowFactor); - return this; - } - /** - * Login thread will sleep until the specified window factor of time - * from last refresh to ticket's expiry has been reached, at which time - * it will try to renew the ticket. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.8 - * Group: security - * - * @param saslKerberosTicketRenewWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslKerberosTicketRenewWindowFactor( - String saslKerberosTicketRenewWindowFactor) { - doSetProperty("saslKerberosTicketRenewWindowFactor", saslKerberosTicketRenewWindowFactor); - return this; - } - /** - * The fully qualified name of a SASL login callback handler class that - * implements the AuthenticateCallbackHandler interface. For brokers, - * login callback handler config must be prefixed with listener prefix - * and SASL mechanism name in lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.login.callback.handler.class=com.example.CustomScramLoginCallbackHandler. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslLoginCallbackHandlerClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginCallbackHandlerClass( - String saslLoginCallbackHandlerClass) { - doSetProperty("saslLoginCallbackHandlerClass", saslLoginCallbackHandlerClass); - return this; - } - /** - * The fully qualified name of a class that implements the Login - * interface. For brokers, login config must be prefixed with listener - * prefix and SASL mechanism name in lower-case. For example, - * listener.name.sasl_ssl.scram-sha-256.sasl.login.class=com.example.CustomScramLogin. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param saslLoginClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginClass(String saslLoginClass) { - doSetProperty("saslLoginClass", saslLoginClass); - return this; - } - /** - * The amount of buffer time before credential expiration to maintain - * when refreshing a credential, in seconds. If a refresh would - * otherwise occur closer to expiration than the number of buffer - * seconds then the refresh will be moved up to maintain as much of the - * buffer time as possible. Legal values are between 0 and 3600 (1 - * hour); a default value of 300 (5 minutes) is used if no value is - * specified. This value and sasl.login.refresh.min.period.seconds are - * both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>short</code> type. - * - * Default: 300 - * Group: security - * - * @param saslLoginRefreshBufferSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginRefreshBufferSeconds( - short saslLoginRefreshBufferSeconds) { - doSetProperty("saslLoginRefreshBufferSeconds", saslLoginRefreshBufferSeconds); - return this; - } - /** - * The amount of buffer time before credential expiration to maintain - * when refreshing a credential, in seconds. If a refresh would - * otherwise occur closer to expiration than the number of buffer - * seconds then the refresh will be moved up to maintain as much of the - * buffer time as possible. Legal values are between 0 and 3600 (1 - * hour); a default value of 300 (5 minutes) is used if no value is - * specified. This value and sasl.login.refresh.min.period.seconds are - * both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>short</code> - * type. - * - * Default: 300 - * Group: security - * - * @param saslLoginRefreshBufferSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginRefreshBufferSeconds( - String saslLoginRefreshBufferSeconds) { - doSetProperty("saslLoginRefreshBufferSeconds", saslLoginRefreshBufferSeconds); - return this; - } - /** - * The desired minimum time for the login refresh thread to wait before - * refreshing a credential, in seconds. Legal values are between 0 and - * 900 (15 minutes); a default value of 60 (1 minute) is used if no - * value is specified. This value and sasl.login.refresh.buffer.seconds - * are both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>short</code> type. - * - * Default: 60 - * Group: security - * - * @param saslLoginRefreshMinPeriodSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginRefreshMinPeriodSeconds( - short saslLoginRefreshMinPeriodSeconds) { - doSetProperty("saslLoginRefreshMinPeriodSeconds", saslLoginRefreshMinPeriodSeconds); - return this; - } - /** - * The desired minimum time for the login refresh thread to wait before - * refreshing a credential, in seconds. Legal values are between 0 and - * 900 (15 minutes); a default value of 60 (1 minute) is used if no - * value is specified. This value and sasl.login.refresh.buffer.seconds - * are both ignored if their sum exceeds the remaining lifetime of a - * credential. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>short</code> - * type. - * - * Default: 60 - * Group: security - * - * @param saslLoginRefreshMinPeriodSeconds the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginRefreshMinPeriodSeconds( - String saslLoginRefreshMinPeriodSeconds) { - doSetProperty("saslLoginRefreshMinPeriodSeconds", saslLoginRefreshMinPeriodSeconds); - return this; - } - /** - * Login refresh thread will sleep until the specified window factor - * relative to the credential's lifetime has been reached, at which time - * it will try to refresh the credential. Legal values are between 0.5 - * (50%) and 1.0 (100%) inclusive; a default value of 0.8 (80%) is used - * if no value is specified. Currently applies only to OAUTHBEARER. - * - * The option is a: <code>double</code> type. - * - * Default: 0.8 - * Group: security - * - * @param saslLoginRefreshWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginRefreshWindowFactor( - double saslLoginRefreshWindowFactor) { - doSetProperty("saslLoginRefreshWindowFactor", saslLoginRefreshWindowFactor); - return this; - } - /** - * Login refresh thread will sleep until the specified window factor - * relative to the credential's lifetime has been reached, at which time - * it will try to refresh the credential. Legal values are between 0.5 - * (50%) and 1.0 (100%) inclusive; a default value of 0.8 (80%) is used - * if no value is specified. Currently applies only to OAUTHBEARER. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.8 - * Group: security - * - * @param saslLoginRefreshWindowFactor the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginRefreshWindowFactor( - String saslLoginRefreshWindowFactor) { - doSetProperty("saslLoginRefreshWindowFactor", saslLoginRefreshWindowFactor); - return this; - } - /** - * The maximum amount of random jitter relative to the credential's - * lifetime that is added to the login refresh thread's sleep time. - * Legal values are between 0 and 0.25 (25%) inclusive; a default value - * of 0.05 (5%) is used if no value is specified. Currently applies only - * to OAUTHBEARER. - * - * The option is a: <code>double</code> type. - * - * Default: 0.05 - * Group: security - * - * @param saslLoginRefreshWindowJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginRefreshWindowJitter( - double saslLoginRefreshWindowJitter) { - doSetProperty("saslLoginRefreshWindowJitter", saslLoginRefreshWindowJitter); - return this; - } - /** - * The maximum amount of random jitter relative to the credential's - * lifetime that is added to the login refresh thread's sleep time. - * Legal values are between 0 and 0.25 (25%) inclusive; a default value - * of 0.05 (5%) is used if no value is specified. Currently applies only - * to OAUTHBEARER. - * - * The option will be converted to a <code>double</code> - * type. - * - * Default: 0.05 - * Group: security - * - * @param saslLoginRefreshWindowJitter the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslLoginRefreshWindowJitter( - String saslLoginRefreshWindowJitter) { - doSetProperty("saslLoginRefreshWindowJitter", saslLoginRefreshWindowJitter); - return this; - } - /** - * SASL mechanism used for client connections. This may be any mechanism - * for which a security provider is available. GSSAPI is the default - * mechanism. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: GSSAPI - * Group: security - * - * @param saslMechanism the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder saslMechanism(String saslMechanism) { - doSetProperty("saslMechanism", saslMechanism); - return this; - } - /** - * Protocol used to communicate with brokers. Valid values are: - * PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: PLAINTEXT - * Group: security - * - * @param securityProtocol the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder securityProtocol( - String securityProtocol) { - doSetProperty("securityProtocol", securityProtocol); - return this; - } - /** - * A list of configurable creator classes each returning a provider - * implementing security algorithms. These classes should implement the - * org.apache.kafka.common.security.auth.SecurityProviderCreator - * interface. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param securityProviders the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder securityProviders( - String securityProviders) { - doSetProperty("securityProviders", securityProviders); - return this; - } - /** - * A list of cipher suites. This is a named combination of - * authentication, encryption, MAC and key exchange algorithm used to - * negotiate the security settings for a network connection using TLS or - * SSL network protocol. By default all the available cipher suites are - * supported. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslCipherSuites the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslCipherSuites(String sslCipherSuites) { - doSetProperty("sslCipherSuites", sslCipherSuites); - return this; - } - /** - * The list of protocols enabled for SSL connections. The default is - * 'TLSv1.2,TLSv1.3' when running with Java 11 or newer, 'TLSv1.2' - * otherwise. With the default value for Java 11, clients and servers - * will prefer TLSv1.3 if both support it and fallback to TLSv1.2 - * otherwise (assuming both support at least TLSv1.2). This default - * should be fine for most cases. Also see the config documentation for - * ssl.protocol. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: TLSv1.2,TLSv1.3 - * Group: security - * - * @param sslEnabledProtocols the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslEnabledProtocols( - String sslEnabledProtocols) { - doSetProperty("sslEnabledProtocols", sslEnabledProtocols); - return this; - } - /** - * The endpoint identification algorithm to validate server hostname - * using server certificate. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: https - * Group: security - * - * @param sslEndpointIdentificationAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslEndpointIdentificationAlgorithm( - String sslEndpointIdentificationAlgorithm) { - doSetProperty("sslEndpointIdentificationAlgorithm", sslEndpointIdentificationAlgorithm); - return this; - } - /** - * The class of type - * org.apache.kafka.common.security.auth.SslEngineFactory to provide - * SSLEngine objects. Default value is - * org.apache.kafka.common.security.ssl.DefaultSslEngineFactory. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslEngineFactoryClass the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslEngineFactoryClass( - String sslEngineFactoryClass) { - doSetProperty("sslEngineFactoryClass", sslEngineFactoryClass); - return this; - } - /** - * The algorithm used by key manager factory for SSL connections. - * Default value is the key manager factory algorithm configured for the - * Java Virtual Machine. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: SunX509 - * Group: security - * - * @param sslKeymanagerAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslKeymanagerAlgorithm( - String sslKeymanagerAlgorithm) { - doSetProperty("sslKeymanagerAlgorithm", sslKeymanagerAlgorithm); - return this; - } - /** - * The password of the private key in the key store file orthe PEM key - * specified in ssl.keystore.key'. This is required for clients only if - * two-way authentication is configured. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeyPassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslKeyPassword(String sslKeyPassword) { - doSetProperty("sslKeyPassword", sslKeyPassword); - return this; - } - /** - * Certificate chain in the format specified by 'ssl.keystore.type'. - * Default SSL engine factory supports only PEM format with a list of - * X.509 certificates. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreCertificateChain the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslKeystoreCertificateChain( - String sslKeystoreCertificateChain) { - doSetProperty("sslKeystoreCertificateChain", sslKeystoreCertificateChain); - return this; - } - /** - * Private key in the format specified by 'ssl.keystore.type'. Default - * SSL engine factory supports only PEM format with PKCS#8 keys. If the - * key is encrypted, key password must be specified using - * 'ssl.key.password'. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreKey the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslKeystoreKey(String sslKeystoreKey) { - doSetProperty("sslKeystoreKey", sslKeystoreKey); - return this; - } - /** - * The location of the key store file. This is optional for client and - * can be used for two-way authentication for client. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystoreLocation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslKeystoreLocation( - String sslKeystoreLocation) { - doSetProperty("sslKeystoreLocation", sslKeystoreLocation); - return this; - } - /** - * The store password for the key store file. This is optional for - * client and only needed if 'ssl.keystore.location' is configured. Key - * store password is not supported for PEM format. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslKeystorePassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslKeystorePassword( - String sslKeystorePassword) { - doSetProperty("sslKeystorePassword", sslKeystorePassword); - return this; - } - /** - * The file format of the key store file. This is optional for client. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: JKS - * Group: security - * - * @param sslKeystoreType the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslKeystoreType(String sslKeystoreType) { - doSetProperty("sslKeystoreType", sslKeystoreType); - return this; - } - /** - * The SSL protocol used to generate the SSLContext. The default is - * 'TLSv1.3' when running with Java 11 or newer, 'TLSv1.2' otherwise. - * This value should be fine for most use cases. Allowed values in - * recent JVMs are 'TLSv1.2' and 'TLSv1.3'. 'TLS', 'TLSv1.1', 'SSL', - * 'SSLv2' and 'SSLv3' may be supported in older JVMs, but their usage - * is discouraged due to known security vulnerabilities. With the - * default value for this config and 'ssl.enabled.protocols', clients - * will downgrade to 'TLSv1.2' if the server does not support 'TLSv1.3'. - * If this config is set to 'TLSv1.2', clients will not use 'TLSv1.3' - * even if it is one of the values in ssl.enabled.protocols and the - * server only supports 'TLSv1.3'. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: TLSv1.2 - * Group: security - * - * @param sslProtocol the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslProtocol(String sslProtocol) { - doSetProperty("sslProtocol", sslProtocol); - return this; - } - /** - * The name of the security provider used for SSL connections. Default - * value is the default security provider of the JVM. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslProvider the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslProvider(String sslProvider) { - doSetProperty("sslProvider", sslProvider); - return this; - } - /** - * The SecureRandom PRNG implementation to use for SSL cryptography - * operations. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslSecureRandomImplementation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslSecureRandomImplementation( - String sslSecureRandomImplementation) { - doSetProperty("sslSecureRandomImplementation", sslSecureRandomImplementation); - return this; - } - /** - * The algorithm used by trust manager factory for SSL connections. - * Default value is the trust manager factory algorithm configured for - * the Java Virtual Machine. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: PKIX - * Group: security - * - * @param sslTrustmanagerAlgorithm the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslTrustmanagerAlgorithm( - String sslTrustmanagerAlgorithm) { - doSetProperty("sslTrustmanagerAlgorithm", sslTrustmanagerAlgorithm); - return this; - } - /** - * Trusted certificates in the format specified by - * 'ssl.truststore.type'. Default SSL engine factory supports only PEM - * format with X.509 certificates. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststoreCertificates the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslTruststoreCertificates( - String sslTruststoreCertificates) { - doSetProperty("sslTruststoreCertificates", sslTruststoreCertificates); - return this; - } - /** - * The location of the trust store file. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststoreLocation the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslTruststoreLocation( - String sslTruststoreLocation) { - doSetProperty("sslTruststoreLocation", sslTruststoreLocation); - return this; - } - /** - * The password for the trust store file. If a password is not set, - * trust store file configured will still be used, but integrity - * checking is disabled. Trust store password is not supported for PEM - * format. - * - * The option is a: <code>java.lang.String</code> type. - * - * Group: security - * - * @param sslTruststorePassword the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslTruststorePassword( - String sslTruststorePassword) { - doSetProperty("sslTruststorePassword", sslTruststorePassword); - return this; - } - /** - * The file format of the trust store file. - * - * The option is a: <code>java.lang.String</code> type. - * - * Default: JKS - * Group: security - * - * @param sslTruststoreType the value to set - * @return the dsl builder - */ - default VertxKafkaEndpointBuilder sslTruststoreType( - String sslTruststoreType) { - doSetProperty("sslTruststoreType", sslTruststoreType); - return this; - } - } - - /** - * Advanced builder for endpoint for the Vert.x Kafka component. - */ - public interface AdvancedVertxKafkaEndpointBuilder - extends - AdvancedVertxKafkaEndpointConsumerBuilder, - AdvancedVertxKafkaEndpointProducerBuilder { - default VertxKafkaEndpointBuilder basic() { - return (VertxKafkaEndpointBuilder) this; - } - } - - public interface VertxKafkaBuilders { - /** - * Vert.x Kafka (camel-vertx-kafka) - * Sent and receive messages to/from an Apache Kafka broker using vert.x - * Kafka client - * - * Category: messaging - * Since: 3.7 - * Maven coordinates: org.apache.camel:camel-vertx-kafka - * - * @return the dsl builder for the headers' name. - */ - default VertxKafkaHeaderNameBuilder vertxKafka() { - return VertxKafkaHeaderNameBuilder.INSTANCE; - } - /** - * Vert.x Kafka (camel-vertx-kafka) - * Sent and receive messages to/from an Apache Kafka broker using vert.x - * Kafka client - * - * Category: messaging - * Since: 3.7 - * Maven coordinates: org.apache.camel:camel-vertx-kafka - * - * Syntax: vertx-kafka:topic - * - * Path parameter: topic (required) - * Name of the topic to use. On the consumer you can use comma to - * separate multiple topics. A producer can only send a message to a - * single topic. - * - * @param path topic - * @return the dsl builder - */ - @Deprecated - default VertxKafkaEndpointBuilder vertxKafka(String path) { - return VertxKafkaEndpointBuilderFactory.endpointBuilder("vertx-kafka", path); - } - /** - * Vert.x Kafka (camel-vertx-kafka) - * Sent and receive messages to/from an Apache Kafka broker using vert.x - * Kafka client - * - * Category: messaging - * Since: 3.7 - * Maven coordinates: org.apache.camel:camel-vertx-kafka - * - * Syntax: vertx-kafka:topic - * - * Path parameter: topic (required) - * Name of the topic to use. On the consumer you can use comma to - * separate multiple topics. A producer can only send a message to a - * single topic. - * - * @param componentName to use a custom component name for the endpoint - * instead of the default name - * @param path topic - * @return the dsl builder - */ - @Deprecated - default VertxKafkaEndpointBuilder vertxKafka( - String componentName, - String path) { - return VertxKafkaEndpointBuilderFactory.endpointBuilder(componentName, path); - } - } - - /** - * The builder of headers' name for the Vert.x Kafka component. - */ - public static class VertxKafkaHeaderNameBuilder { - /** - * The internal instance of the builder used to access to all the - * methods representing the name of headers. - */ - private static final VertxKafkaHeaderNameBuilder INSTANCE = new VertxKafkaHeaderNameBuilder(); - - /** - * Producer: Explicitly specify the partition identifier, for example - * partition 0. This will trigger the component to produce all the - * massages to the specified partition. Consumer: The partition - * identifier where the message were consumed from. - * - * The option is a: {@code Integer} type. - * - * Group: common - * - * @return the name of the header {@code VertxKafkaPartitionId}. - */ - public String vertxKafkaPartitionId() { - return "VertxKafkaPartitionId"; - } - - /** - * Producer: Explicitly specify the message key, if partition ID is not - * specified, this will trigger the messages to go into the same - * partition. Consumer: The message key. - * - * The option is a: {@code String} type. - * - * Group: common - * - * @return the name of the header {@code VertxKafkaMessageKey}. - */ - public String vertxKafkaMessageKey() { - return "VertxKafkaMessageKey"; - } - - /** - * Producer: Explicitly specify the topic to where produce the messages, - * this will be preserved in case of header aggregation. Consumer: The - * topic from where the message originated. - * - * The option is a: {@code String} type. - * - * Group: common - * - * @return the name of the header {@code VertxKafkaTopic}. - */ - public String vertxKafkaTopic() { - return "VertxKafkaTopic"; - } - - /** - * Produced record metadata. - * - * The option is a: {@code List} type. - * - * Group: producer - * - * @return the name of the header {@code VertxKafkaRecordMetadata}. - */ - public String vertxKafkaRecordMetadata() { - return "VertxKafkaRecordMetadata"; - } - - /** - * The offset of the message in Kafka topic. - * - * The option is a: {@code Long} type. - * - * Group: consumer - * - * @return the name of the header {@code VertxKafkaOffset}. - */ - public String vertxKafkaOffset() { - return "VertxKafkaOffset"; - } - - /** - * The record Kafka headers. - * - * The option is a: {@code List} type. - * - * Group: consumer - * - * @return the name of the header {@code VertxKafkaHeaders}. - */ - public String vertxKafkaHeaders() { - return "VertxKafkaHeaders"; - } - - /** - * The timestamp of this record. - * - * The option is a: {@code Long} type. - * - * Group: consumer - * - * @return the name of the header {@code VertxKafkaTimestamp}. - */ - public String vertxKafkaTimestamp() { - return "VertxKafkaTimestamp"; - } - - /** - * The ProducerRecord also has an associated timestamp. If the user did - * provide a timestamp, the producer will stamp the record with the - * provided timestamp and the header is not preserved. - * - * The option is a: {@code Long} type. - * - * Group: producer - * - * @return the name of the header {@code VertxKafkaOverrideTimestamp}. - */ - public String vertxKafkaOverrideTimestamp() { - return "VertxKafkaOverrideTimestamp"; - } - - /** - * Explicitly specify the topic to where produce the messages, this will - * not be preserved in case of header aggregation and it will take - * precedence over CamelVertxKafkaTopic. - * - * The option is a: {@code String} type. - * - * Group: producer - * - * @return the name of the header {@code VertxKafkaOverrideTopic}. - */ - public String vertxKafkaOverrideTopic() { - return "VertxKafkaOverrideTopic"; - } - } - @Deprecated - static VertxKafkaEndpointBuilder endpointBuilder( - String componentName, - String path) { - class VertxKafkaEndpointBuilderImpl extends AbstractEndpointBuilder implements VertxKafkaEndpointBuilder, AdvancedVertxKafkaEndpointBuilder { - public VertxKafkaEndpointBuilderImpl(String path) { - super(componentName, path); - } - } - return new VertxKafkaEndpointBuilderImpl(path); - } -} \ No newline at end of file diff --git a/parent/pom.xml b/parent/pom.xml index 4ddb39db2cce6..d499fc21043ae 100644 --- a/parent/pom.xml +++ b/parent/pom.xml @@ -2732,11 +2732,6 @@ camel-servicenow-maven-plugin ${project.version} - - org.apache.camel.maven - camel-vertx-kafka-maven-plugin - ${project.version} -