Skip to content
Permalink
Branch: master
Find file Copy path
Find file Copy path
23 contributors

Users who have contributed to this file

@edenhill @rondagostino @rthalley @mageshn @HeChuanXUPT @turtlesoupy @tbsaunde @skarlsson @sczizzo @paravoid @noahdav @benesch @mhowlett @NZKoz @julien-lecomte @hqin @fede1024 @eugpermar @lambdaknight @edoardocomar @johnistan @briot @barrotsteindev
5948 lines (5201 sloc) 200 KB
/*
* librdkafka - Apache Kafka C library
*
* Copyright (c) 2012-2018 Magnus Edenhill
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
*
* 1. Redistributions of source code must retain the above copyright notice,
* this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright notice,
* this list of conditions and the following disclaimer in the documentation
* and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
/**
* @file rdkafka.h
* @brief Apache Kafka C/C++ consumer and producer client library.
*
* rdkafka.h contains the public API for librdkafka.
* The API is documented in this file as comments prefixing the function, type,
* enum, define, etc.
*
* @sa For the C++ interface see rdkafkacpp.h
*
* @tableofcontents
*/
/* @cond NO_DOC */
#ifndef _RDKAFKA_H_
#define _RDKAFKA_H_
#include <stdio.h>
#include <inttypes.h>
#include <sys/types.h>
#ifdef __cplusplus
extern "C" {
#if 0
} /* Restore indent */
#endif
#endif
#ifdef _MSC_VER
#include <basetsd.h>
#ifndef WIN32_MEAN_AND_LEAN
#define WIN32_MEAN_AND_LEAN
#endif
#include <Winsock2.h> /* for sockaddr, .. */
typedef SSIZE_T ssize_t;
#define RD_UNUSED
#define RD_INLINE __inline
#define RD_DEPRECATED __declspec(deprecated)
#undef RD_EXPORT
#ifdef LIBRDKAFKA_STATICLIB
#define RD_EXPORT
#else
#ifdef LIBRDKAFKA_EXPORTS
#define RD_EXPORT __declspec(dllexport)
#else
#define RD_EXPORT __declspec(dllimport)
#endif
#ifndef LIBRDKAFKA_TYPECHECKS
#define LIBRDKAFKA_TYPECHECKS 0
#endif
#endif
#else
#include <sys/socket.h> /* for sockaddr, .. */
#define RD_UNUSED __attribute__((unused))
#define RD_INLINE inline
#define RD_EXPORT
#define RD_DEPRECATED __attribute__((deprecated))
#ifndef LIBRDKAFKA_TYPECHECKS
#define LIBRDKAFKA_TYPECHECKS 1
#endif
#endif
/**
* @brief Type-checking macros
* Compile-time checking that \p ARG is of type \p TYPE.
* @returns \p RET
*/
#if LIBRDKAFKA_TYPECHECKS
#define _LRK_TYPECHECK(RET,TYPE,ARG) \
({ if (0) { TYPE __t RD_UNUSED = (ARG); } RET; })
#define _LRK_TYPECHECK2(RET,TYPE,ARG,TYPE2,ARG2) \
({ \
if (0) { \
TYPE __t RD_UNUSED = (ARG); \
TYPE2 __t2 RD_UNUSED = (ARG2); \
} \
RET; })
#define _LRK_TYPECHECK3(RET,TYPE,ARG,TYPE2,ARG2,TYPE3,ARG3) \
({ \
if (0) { \
TYPE __t RD_UNUSED = (ARG); \
TYPE2 __t2 RD_UNUSED = (ARG2); \
TYPE3 __t3 RD_UNUSED = (ARG3); \
} \
RET; })
#else
#define _LRK_TYPECHECK(RET,TYPE,ARG) (RET)
#define _LRK_TYPECHECK2(RET,TYPE,ARG,TYPE2,ARG2) (RET)
#define _LRK_TYPECHECK3(RET,TYPE,ARG,TYPE2,ARG2,TYPE3,ARG3) (RET)
#endif
/* @endcond */
/**
* @name librdkafka version
* @{
*
*
*/
/**
* @brief librdkafka version
*
* Interpreted as hex \c MM.mm.rr.xx:
* - MM = Major
* - mm = minor
* - rr = revision
* - xx = pre-release id (0xff is the final release)
*
* E.g.: \c 0x000801ff = 0.8.1
*
* @remark This value should only be used during compile time,
* for runtime checks of version use rd_kafka_version()
*/
#define RD_KAFKA_VERSION 0x010100ff
/**
* @brief Returns the librdkafka version as integer.
*
* @returns Version integer.
*
* @sa See RD_KAFKA_VERSION for how to parse the integer format.
* @sa Use rd_kafka_version_str() to retreive the version as a string.
*/
RD_EXPORT
int rd_kafka_version(void);
/**
* @brief Returns the librdkafka version as string.
*
* @returns Version string
*/
RD_EXPORT
const char *rd_kafka_version_str (void);
/**@}*/
/**
* @name Constants, errors, types
* @{
*
*
*/
/**
* @enum rd_kafka_type_t
*
* @brief rd_kafka_t handle type.
*
* @sa rd_kafka_new()
*/
typedef enum rd_kafka_type_t {
RD_KAFKA_PRODUCER, /**< Producer client */
RD_KAFKA_CONSUMER /**< Consumer client */
} rd_kafka_type_t;
/**
* @enum Timestamp types
*
* @sa rd_kafka_message_timestamp()
*/
typedef enum rd_kafka_timestamp_type_t {
RD_KAFKA_TIMESTAMP_NOT_AVAILABLE, /**< Timestamp not available */
RD_KAFKA_TIMESTAMP_CREATE_TIME, /**< Message creation time */
RD_KAFKA_TIMESTAMP_LOG_APPEND_TIME /**< Log append time */
} rd_kafka_timestamp_type_t;
/**
* @brief Retrieve supported debug contexts for use with the \c \"debug\"
* configuration property. (runtime)
*
* @returns Comma-separated list of available debugging contexts.
*/
RD_EXPORT
const char *rd_kafka_get_debug_contexts(void);
/**
* @brief Supported debug contexts. (compile time)
*
* @deprecated This compile time value may be outdated at runtime due to
* linking another version of the library.
* Use rd_kafka_get_debug_contexts() instead.
*/
#define RD_KAFKA_DEBUG_CONTEXTS \
"all,generic,broker,topic,metadata,feature,queue,msg,protocol,cgrp,security,fetch,interceptor,plugin,consumer,admin,eos"
/* @cond NO_DOC */
/* Private types to provide ABI compatibility */
typedef struct rd_kafka_s rd_kafka_t;
typedef struct rd_kafka_topic_s rd_kafka_topic_t;
typedef struct rd_kafka_conf_s rd_kafka_conf_t;
typedef struct rd_kafka_topic_conf_s rd_kafka_topic_conf_t;
typedef struct rd_kafka_queue_s rd_kafka_queue_t;
typedef struct rd_kafka_op_s rd_kafka_event_t;
typedef struct rd_kafka_topic_result_s rd_kafka_topic_result_t;
/* @endcond */
/**
* @enum rd_kafka_resp_err_t
* @brief Error codes.
*
* The negative error codes delimited by two underscores
* (\c RD_KAFKA_RESP_ERR__..) denotes errors internal to librdkafka and are
* displayed as \c \"Local: \<error string..\>\", while the error codes
* delimited by a single underscore (\c RD_KAFKA_RESP_ERR_..) denote broker
* errors and are displayed as \c \"Broker: \<error string..\>\".
*
* @sa Use rd_kafka_err2str() to translate an error code a human readable string
*/
typedef enum {
/* Internal errors to rdkafka: */
/** Begin internal error codes */
RD_KAFKA_RESP_ERR__BEGIN = -200,
/** Received message is incorrect */
RD_KAFKA_RESP_ERR__BAD_MSG = -199,
/** Bad/unknown compression */
RD_KAFKA_RESP_ERR__BAD_COMPRESSION = -198,
/** Broker is going away */
RD_KAFKA_RESP_ERR__DESTROY = -197,
/** Generic failure */
RD_KAFKA_RESP_ERR__FAIL = -196,
/** Broker transport failure */
RD_KAFKA_RESP_ERR__TRANSPORT = -195,
/** Critical system resource */
RD_KAFKA_RESP_ERR__CRIT_SYS_RESOURCE = -194,
/** Failed to resolve broker */
RD_KAFKA_RESP_ERR__RESOLVE = -193,
/** Produced message timed out*/
RD_KAFKA_RESP_ERR__MSG_TIMED_OUT = -192,
/** Reached the end of the topic+partition queue on
* the broker. Not really an error. */
RD_KAFKA_RESP_ERR__PARTITION_EOF = -191,
/** Permanent: Partition does not exist in cluster. */
RD_KAFKA_RESP_ERR__UNKNOWN_PARTITION = -190,
/** File or filesystem error */
RD_KAFKA_RESP_ERR__FS = -189,
/** Permanent: Topic does not exist in cluster. */
RD_KAFKA_RESP_ERR__UNKNOWN_TOPIC = -188,
/** All broker connections are down. */
RD_KAFKA_RESP_ERR__ALL_BROKERS_DOWN = -187,
/** Invalid argument, or invalid configuration */
RD_KAFKA_RESP_ERR__INVALID_ARG = -186,
/** Operation timed out */
RD_KAFKA_RESP_ERR__TIMED_OUT = -185,
/** Queue is full */
RD_KAFKA_RESP_ERR__QUEUE_FULL = -184,
/** ISR count < required.acks */
RD_KAFKA_RESP_ERR__ISR_INSUFF = -183,
/** Broker node update */
RD_KAFKA_RESP_ERR__NODE_UPDATE = -182,
/** SSL error */
RD_KAFKA_RESP_ERR__SSL = -181,
/** Waiting for coordinator to become available. */
RD_KAFKA_RESP_ERR__WAIT_COORD = -180,
/** Unknown client group */
RD_KAFKA_RESP_ERR__UNKNOWN_GROUP = -179,
/** Operation in progress */
RD_KAFKA_RESP_ERR__IN_PROGRESS = -178,
/** Previous operation in progress, wait for it to finish. */
RD_KAFKA_RESP_ERR__PREV_IN_PROGRESS = -177,
/** This operation would interfere with an existing subscription */
RD_KAFKA_RESP_ERR__EXISTING_SUBSCRIPTION = -176,
/** Assigned partitions (rebalance_cb) */
RD_KAFKA_RESP_ERR__ASSIGN_PARTITIONS = -175,
/** Revoked partitions (rebalance_cb) */
RD_KAFKA_RESP_ERR__REVOKE_PARTITIONS = -174,
/** Conflicting use */
RD_KAFKA_RESP_ERR__CONFLICT = -173,
/** Wrong state */
RD_KAFKA_RESP_ERR__STATE = -172,
/** Unknown protocol */
RD_KAFKA_RESP_ERR__UNKNOWN_PROTOCOL = -171,
/** Not implemented */
RD_KAFKA_RESP_ERR__NOT_IMPLEMENTED = -170,
/** Authentication failure*/
RD_KAFKA_RESP_ERR__AUTHENTICATION = -169,
/** No stored offset */
RD_KAFKA_RESP_ERR__NO_OFFSET = -168,
/** Outdated */
RD_KAFKA_RESP_ERR__OUTDATED = -167,
/** Timed out in queue */
RD_KAFKA_RESP_ERR__TIMED_OUT_QUEUE = -166,
/** Feature not supported by broker */
RD_KAFKA_RESP_ERR__UNSUPPORTED_FEATURE = -165,
/** Awaiting cache update */
RD_KAFKA_RESP_ERR__WAIT_CACHE = -164,
/** Operation interrupted (e.g., due to yield)) */
RD_KAFKA_RESP_ERR__INTR = -163,
/** Key serialization error */
RD_KAFKA_RESP_ERR__KEY_SERIALIZATION = -162,
/** Value serialization error */
RD_KAFKA_RESP_ERR__VALUE_SERIALIZATION = -161,
/** Key deserialization error */
RD_KAFKA_RESP_ERR__KEY_DESERIALIZATION = -160,
/** Value deserialization error */
RD_KAFKA_RESP_ERR__VALUE_DESERIALIZATION = -159,
/** Partial response */
RD_KAFKA_RESP_ERR__PARTIAL = -158,
/** Modification attempted on read-only object */
RD_KAFKA_RESP_ERR__READ_ONLY = -157,
/** No such entry / item not found */
RD_KAFKA_RESP_ERR__NOENT = -156,
/** Read underflow */
RD_KAFKA_RESP_ERR__UNDERFLOW = -155,
/** Invalid type */
RD_KAFKA_RESP_ERR__INVALID_TYPE = -154,
/** Retry operation */
RD_KAFKA_RESP_ERR__RETRY = -153,
/** Purged in queue */
RD_KAFKA_RESP_ERR__PURGE_QUEUE = -152,
/** Purged in flight */
RD_KAFKA_RESP_ERR__PURGE_INFLIGHT = -151,
/** Fatal error: see rd_kafka_fatal_error() */
RD_KAFKA_RESP_ERR__FATAL = -150,
/** Inconsistent state */
RD_KAFKA_RESP_ERR__INCONSISTENT = -149,
/** Gap-less ordering would not be guaranteed if proceeding */
RD_KAFKA_RESP_ERR__GAPLESS_GUARANTEE = -148,
/** Maximum poll interval exceeded */
RD_KAFKA_RESP_ERR__MAX_POLL_EXCEEDED = -147,
/** End internal error codes */
RD_KAFKA_RESP_ERR__END = -100,
/* Kafka broker errors: */
/** Unknown broker error */
RD_KAFKA_RESP_ERR_UNKNOWN = -1,
/** Success */
RD_KAFKA_RESP_ERR_NO_ERROR = 0,
/** Offset out of range */
RD_KAFKA_RESP_ERR_OFFSET_OUT_OF_RANGE = 1,
/** Invalid message */
RD_KAFKA_RESP_ERR_INVALID_MSG = 2,
/** Unknown topic or partition */
RD_KAFKA_RESP_ERR_UNKNOWN_TOPIC_OR_PART = 3,
/** Invalid message size */
RD_KAFKA_RESP_ERR_INVALID_MSG_SIZE = 4,
/** Leader not available */
RD_KAFKA_RESP_ERR_LEADER_NOT_AVAILABLE = 5,
/** Not leader for partition */
RD_KAFKA_RESP_ERR_NOT_LEADER_FOR_PARTITION = 6,
/** Request timed out */
RD_KAFKA_RESP_ERR_REQUEST_TIMED_OUT = 7,
/** Broker not available */
RD_KAFKA_RESP_ERR_BROKER_NOT_AVAILABLE = 8,
/** Replica not available */
RD_KAFKA_RESP_ERR_REPLICA_NOT_AVAILABLE = 9,
/** Message size too large */
RD_KAFKA_RESP_ERR_MSG_SIZE_TOO_LARGE = 10,
/** StaleControllerEpochCode */
RD_KAFKA_RESP_ERR_STALE_CTRL_EPOCH = 11,
/** Offset metadata string too large */
RD_KAFKA_RESP_ERR_OFFSET_METADATA_TOO_LARGE = 12,
/** Broker disconnected before response received */
RD_KAFKA_RESP_ERR_NETWORK_EXCEPTION = 13,
/** Group coordinator load in progress */
RD_KAFKA_RESP_ERR_GROUP_LOAD_IN_PROGRESS = 14,
/** Group coordinator not available */
RD_KAFKA_RESP_ERR_GROUP_COORDINATOR_NOT_AVAILABLE = 15,
/** Not coordinator for group */
RD_KAFKA_RESP_ERR_NOT_COORDINATOR_FOR_GROUP = 16,
/** Invalid topic */
RD_KAFKA_RESP_ERR_TOPIC_EXCEPTION = 17,
/** Message batch larger than configured server segment size */
RD_KAFKA_RESP_ERR_RECORD_LIST_TOO_LARGE = 18,
/** Not enough in-sync replicas */
RD_KAFKA_RESP_ERR_NOT_ENOUGH_REPLICAS = 19,
/** Message(s) written to insufficient number of in-sync replicas */
RD_KAFKA_RESP_ERR_NOT_ENOUGH_REPLICAS_AFTER_APPEND = 20,
/** Invalid required acks value */
RD_KAFKA_RESP_ERR_INVALID_REQUIRED_ACKS = 21,
/** Specified group generation id is not valid */
RD_KAFKA_RESP_ERR_ILLEGAL_GENERATION = 22,
/** Inconsistent group protocol */
RD_KAFKA_RESP_ERR_INCONSISTENT_GROUP_PROTOCOL = 23,
/** Invalid group.id */
RD_KAFKA_RESP_ERR_INVALID_GROUP_ID = 24,
/** Unknown member */
RD_KAFKA_RESP_ERR_UNKNOWN_MEMBER_ID = 25,
/** Invalid session timeout */
RD_KAFKA_RESP_ERR_INVALID_SESSION_TIMEOUT = 26,
/** Group rebalance in progress */
RD_KAFKA_RESP_ERR_REBALANCE_IN_PROGRESS = 27,
/** Commit offset data size is not valid */
RD_KAFKA_RESP_ERR_INVALID_COMMIT_OFFSET_SIZE = 28,
/** Topic authorization failed */
RD_KAFKA_RESP_ERR_TOPIC_AUTHORIZATION_FAILED = 29,
/** Group authorization failed */
RD_KAFKA_RESP_ERR_GROUP_AUTHORIZATION_FAILED = 30,
/** Cluster authorization failed */
RD_KAFKA_RESP_ERR_CLUSTER_AUTHORIZATION_FAILED = 31,
/** Invalid timestamp */
RD_KAFKA_RESP_ERR_INVALID_TIMESTAMP = 32,
/** Unsupported SASL mechanism */
RD_KAFKA_RESP_ERR_UNSUPPORTED_SASL_MECHANISM = 33,
/** Illegal SASL state */
RD_KAFKA_RESP_ERR_ILLEGAL_SASL_STATE = 34,
/** Unuspported version */
RD_KAFKA_RESP_ERR_UNSUPPORTED_VERSION = 35,
/** Topic already exists */
RD_KAFKA_RESP_ERR_TOPIC_ALREADY_EXISTS = 36,
/** Invalid number of partitions */
RD_KAFKA_RESP_ERR_INVALID_PARTITIONS = 37,
/** Invalid replication factor */
RD_KAFKA_RESP_ERR_INVALID_REPLICATION_FACTOR = 38,
/** Invalid replica assignment */
RD_KAFKA_RESP_ERR_INVALID_REPLICA_ASSIGNMENT = 39,
/** Invalid config */
RD_KAFKA_RESP_ERR_INVALID_CONFIG = 40,
/** Not controller for cluster */
RD_KAFKA_RESP_ERR_NOT_CONTROLLER = 41,
/** Invalid request */
RD_KAFKA_RESP_ERR_INVALID_REQUEST = 42,
/** Message format on broker does not support request */
RD_KAFKA_RESP_ERR_UNSUPPORTED_FOR_MESSAGE_FORMAT = 43,
/** Policy violation */
RD_KAFKA_RESP_ERR_POLICY_VIOLATION = 44,
/** Broker received an out of order sequence number */
RD_KAFKA_RESP_ERR_OUT_OF_ORDER_SEQUENCE_NUMBER = 45,
/** Broker received a duplicate sequence number */
RD_KAFKA_RESP_ERR_DUPLICATE_SEQUENCE_NUMBER = 46,
/** Producer attempted an operation with an old epoch */
RD_KAFKA_RESP_ERR_INVALID_PRODUCER_EPOCH = 47,
/** Producer attempted a transactional operation in an invalid state */
RD_KAFKA_RESP_ERR_INVALID_TXN_STATE = 48,
/** Producer attempted to use a producer id which is not
* currently assigned to its transactional id */
RD_KAFKA_RESP_ERR_INVALID_PRODUCER_ID_MAPPING = 49,
/** Transaction timeout is larger than the maximum
* value allowed by the broker's max.transaction.timeout.ms */
RD_KAFKA_RESP_ERR_INVALID_TRANSACTION_TIMEOUT = 50,
/** Producer attempted to update a transaction while another
* concurrent operation on the same transaction was ongoing */
RD_KAFKA_RESP_ERR_CONCURRENT_TRANSACTIONS = 51,
/** Indicates that the transaction coordinator sending a
* WriteTxnMarker is no longer the current coordinator for a
* given producer */
RD_KAFKA_RESP_ERR_TRANSACTION_COORDINATOR_FENCED = 52,
/** Transactional Id authorization failed */
RD_KAFKA_RESP_ERR_TRANSACTIONAL_ID_AUTHORIZATION_FAILED = 53,
/** Security features are disabled */
RD_KAFKA_RESP_ERR_SECURITY_DISABLED = 54,
/** Operation not attempted */
RD_KAFKA_RESP_ERR_OPERATION_NOT_ATTEMPTED = 55,
/** Disk error when trying to access log file on the disk */
RD_KAFKA_RESP_ERR_KAFKA_STORAGE_ERROR = 56,
/** The user-specified log directory is not found in the broker config */
RD_KAFKA_RESP_ERR_LOG_DIR_NOT_FOUND = 57,
/** SASL Authentication failed */
RD_KAFKA_RESP_ERR_SASL_AUTHENTICATION_FAILED = 58,
/** Unknown Producer Id */
RD_KAFKA_RESP_ERR_UNKNOWN_PRODUCER_ID = 59,
/** Partition reassignment is in progress */
RD_KAFKA_RESP_ERR_REASSIGNMENT_IN_PROGRESS = 60,
/** Delegation Token feature is not enabled */
RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_AUTH_DISABLED = 61,
/** Delegation Token is not found on server */
RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_NOT_FOUND = 62,
/** Specified Principal is not valid Owner/Renewer */
RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_OWNER_MISMATCH = 63,
/** Delegation Token requests are not allowed on this connection */
RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_REQUEST_NOT_ALLOWED = 64,
/** Delegation Token authorization failed */
RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_AUTHORIZATION_FAILED = 65,
/** Delegation Token is expired */
RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_EXPIRED = 66,
/** Supplied principalType is not supported */
RD_KAFKA_RESP_ERR_INVALID_PRINCIPAL_TYPE = 67,
/** The group is not empty */
RD_KAFKA_RESP_ERR_NON_EMPTY_GROUP = 68,
/** The group id does not exist */
RD_KAFKA_RESP_ERR_GROUP_ID_NOT_FOUND = 69,
/** The fetch session ID was not found */
RD_KAFKA_RESP_ERR_FETCH_SESSION_ID_NOT_FOUND = 70,
/** The fetch session epoch is invalid */
RD_KAFKA_RESP_ERR_INVALID_FETCH_SESSION_EPOCH = 71,
/** No matching listener */
RD_KAFKA_RESP_ERR_LISTENER_NOT_FOUND = 72,
/** Topic deletion is disabled */
RD_KAFKA_RESP_ERR_TOPIC_DELETION_DISABLED = 73,
/** Leader epoch is older than broker epoch */
RD_KAFKA_RESP_ERR_FENCED_LEADER_EPOCH = 74,
/** Leader epoch is newer than broker epoch */
RD_KAFKA_RESP_ERR_UNKNOWN_LEADER_EPOCH = 75,
/** Unsupported compression type */
RD_KAFKA_RESP_ERR_UNSUPPORTED_COMPRESSION_TYPE = 76,
/** Broker epoch has changed */
RD_KAFKA_RESP_ERR_STALE_BROKER_EPOCH = 77,
/** Leader high watermark is not caught up */
RD_KAFKA_RESP_ERR_OFFSET_NOT_AVAILABLE = 78,
/** Group member needs a valid member ID */
RD_KAFKA_RESP_ERR_MEMBER_ID_REQUIRED = 79,
/** Preferred leader was not available */
RD_KAFKA_RESP_ERR_PREFERRED_LEADER_NOT_AVAILABLE = 80,
/** Consumer group has reached maximum size */
RD_KAFKA_RESP_ERR_GROUP_MAX_SIZE_REACHED = 81,
RD_KAFKA_RESP_ERR_END_ALL,
} rd_kafka_resp_err_t;
/**
* @brief Error code value, name and description.
* Typically for use with language bindings to automatically expose
* the full set of librdkafka error codes.
*/
struct rd_kafka_err_desc {
rd_kafka_resp_err_t code;/**< Error code */
const char *name; /**< Error name, same as code enum sans prefix */
const char *desc; /**< Human readable error description. */
};
/**
* @brief Returns the full list of error codes.
*/
RD_EXPORT
void rd_kafka_get_err_descs (const struct rd_kafka_err_desc **errdescs,
size_t *cntp);
/**
* @brief Returns a human readable representation of a kafka error.
*
* @param err Error code to translate
*/
RD_EXPORT
const char *rd_kafka_err2str (rd_kafka_resp_err_t err);
/**
* @brief Returns the error code name (enum name).
*
* @param err Error code to translate
*/
RD_EXPORT
const char *rd_kafka_err2name (rd_kafka_resp_err_t err);
/**
* @brief Returns the last error code generated by a legacy API call
* in the current thread.
*
* The legacy APIs are the ones using errno to propagate error value, namely:
* - rd_kafka_topic_new()
* - rd_kafka_consume_start()
* - rd_kafka_consume_stop()
* - rd_kafka_consume()
* - rd_kafka_consume_batch()
* - rd_kafka_consume_callback()
* - rd_kafka_consume_queue()
* - rd_kafka_produce()
*
* The main use for this function is to avoid converting system \p errno
* values to rd_kafka_resp_err_t codes for legacy APIs.
*
* @remark The last error is stored per-thread, if multiple rd_kafka_t handles
* are used in the same application thread the developer needs to
* make sure rd_kafka_last_error() is called immediately after
* a failed API call.
*
* @remark errno propagation from librdkafka is not safe on Windows
* and should not be used, use rd_kafka_last_error() instead.
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_last_error (void);
/**
* @brief Converts the system errno value \p errnox to a rd_kafka_resp_err_t
* error code upon failure from the following functions:
* - rd_kafka_topic_new()
* - rd_kafka_consume_start()
* - rd_kafka_consume_stop()
* - rd_kafka_consume()
* - rd_kafka_consume_batch()
* - rd_kafka_consume_callback()
* - rd_kafka_consume_queue()
* - rd_kafka_produce()
*
* @param errnox System errno value to convert
*
* @returns Appropriate error code for \p errnox
*
* @remark A better alternative is to call rd_kafka_last_error() immediately
* after any of the above functions return -1 or NULL.
*
* @deprecated Use rd_kafka_last_error() to retrieve the last error code
* set by the legacy librdkafka APIs.
*
* @sa rd_kafka_last_error()
*/
RD_EXPORT RD_DEPRECATED
rd_kafka_resp_err_t rd_kafka_errno2err(int errnox);
/**
* @brief Returns the thread-local system errno
*
* On most platforms this is the same as \p errno but in case of different
* runtimes between library and application (e.g., Windows static DLLs)
* this provides a means for exposing the errno librdkafka uses.
*
* @remark The value is local to the current calling thread.
*
* @deprecated Use rd_kafka_last_error() to retrieve the last error code
* set by the legacy librdkafka APIs.
*/
RD_EXPORT RD_DEPRECATED
int rd_kafka_errno (void);
/**
* @brief Returns the first fatal error set on this client instance,
* or RD_KAFKA_RESP_ERR_NO_ERROR if no fatal error has occurred.
*
* This function is to be used with the Idempotent Producer and \c error_cb
* to detect fatal errors.
*
* Generally all errors raised by \c error_cb are to be considered
* informational and temporary, the client will try to recover from all
* errors in a graceful fashion (by retrying, etc).
*
* However, some errors should logically be considered fatal to retain
* consistency; in particular a set of errors that may occur when using the
* Idempotent Producer and the in-order or exactly-once producer guarantees
* can't be satisfied.
*
* @param errstr A human readable error string (nul-terminated) is written to
* this location that must be of at least \p errstr_size bytes.
* The \p errstr is only written to if there is a fatal error.
*
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR if no fatal error has been raised, else
* any other error code.
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_fatal_error (rd_kafka_t *rk,
char *errstr, size_t errstr_size);
/**
* @brief Trigger a fatal error for testing purposes.
*
* Since there is no practical way to trigger real fatal errors in the
* idempotent producer, this method allows an application to trigger
* fabricated fatal errors in tests to check its error handling code.
*
* @param err The underlying error code.
* @param reason A human readable error reason.
* Will be prefixed with "test_fatal_error: " to differentiate
* from real fatal errors.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR if a fatal error was triggered, or
* RD_KAFKA_RESP_ERR__PREV_IN_PROGRESS if a previous fatal error
* has already been triggered.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_test_fatal_error (rd_kafka_t *rk, rd_kafka_resp_err_t err,
const char *reason);
/**
* @brief Topic+Partition place holder
*
* Generic place holder for a Topic+Partition and its related information
* used for multiple purposes:
* - consumer offset (see rd_kafka_commit(), et.al.)
* - group rebalancing callback (rd_kafka_conf_set_rebalance_cb())
* - offset commit result callback (rd_kafka_conf_set_offset_commit_cb())
*/
/**
* @brief Generic place holder for a specific Topic+Partition.
*
* @sa rd_kafka_topic_partition_list_new()
*/
typedef struct rd_kafka_topic_partition_s {
char *topic; /**< Topic name */
int32_t partition; /**< Partition */
int64_t offset; /**< Offset */
void *metadata; /**< Metadata */
size_t metadata_size; /**< Metadata size */
void *opaque; /**< Application opaque */
rd_kafka_resp_err_t err; /**< Error code, depending on use. */
void *_private; /**< INTERNAL USE ONLY,
* INITIALIZE TO ZERO, DO NOT TOUCH */
} rd_kafka_topic_partition_t;
/**
* @brief Destroy a rd_kafka_topic_partition_t.
* @remark This must not be called for elements in a topic partition list.
*/
RD_EXPORT
void rd_kafka_topic_partition_destroy (rd_kafka_topic_partition_t *rktpar);
/**
* @brief A growable list of Topic+Partitions.
*
*/
typedef struct rd_kafka_topic_partition_list_s {
int cnt; /**< Current number of elements */
int size; /**< Current allocated size */
rd_kafka_topic_partition_t *elems; /**< Element array[] */
} rd_kafka_topic_partition_list_t;
/**
* @brief Create a new list/vector Topic+Partition container.
*
* @param size Initial allocated size used when the expected number of
* elements is known or can be estimated.
* Avoids reallocation and possibly relocation of the
* elems array.
*
* @returns A newly allocated Topic+Partition list.
*
* @remark Use rd_kafka_topic_partition_list_destroy() to free all resources
* in use by a list and the list itself.
* @sa rd_kafka_topic_partition_list_add()
*/
RD_EXPORT
rd_kafka_topic_partition_list_t *rd_kafka_topic_partition_list_new (int size);
/**
* @brief Free all resources used by the list and the list itself.
*/
RD_EXPORT
void
rd_kafka_topic_partition_list_destroy (rd_kafka_topic_partition_list_t *rkparlist);
/**
* @brief Add topic+partition to list
*
* @param rktparlist List to extend
* @param topic Topic name (copied)
* @param partition Partition id
*
* @returns The object which can be used to fill in additionals fields.
*/
RD_EXPORT
rd_kafka_topic_partition_t *
rd_kafka_topic_partition_list_add (rd_kafka_topic_partition_list_t *rktparlist,
const char *topic, int32_t partition);
/**
* @brief Add range of partitions from \p start to \p stop inclusive.
*
* @param rktparlist List to extend
* @param topic Topic name (copied)
* @param start Start partition of range
* @param stop Last partition of range (inclusive)
*/
RD_EXPORT
void
rd_kafka_topic_partition_list_add_range (rd_kafka_topic_partition_list_t
*rktparlist,
const char *topic,
int32_t start, int32_t stop);
/**
* @brief Delete partition from list.
*
* @param rktparlist List to modify
* @param topic Topic name to match
* @param partition Partition to match
*
* @returns 1 if partition was found (and removed), else 0.
*
* @remark Any held indices to elems[] are unusable after this call returns 1.
*/
RD_EXPORT
int
rd_kafka_topic_partition_list_del (rd_kafka_topic_partition_list_t *rktparlist,
const char *topic, int32_t partition);
/**
* @brief Delete partition from list by elems[] index.
*
* @returns 1 if partition was found (and removed), else 0.
*
* @sa rd_kafka_topic_partition_list_del()
*/
RD_EXPORT
int
rd_kafka_topic_partition_list_del_by_idx (
rd_kafka_topic_partition_list_t *rktparlist,
int idx);
/**
* @brief Make a copy of an existing list.
*
* @param src The existing list to copy.
*
* @returns A new list fully populated to be identical to \p src
*/
RD_EXPORT
rd_kafka_topic_partition_list_t *
rd_kafka_topic_partition_list_copy (const rd_kafka_topic_partition_list_t *src);
/**
* @brief Set offset to \p offset for \p topic and \p partition
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success or
* RD_KAFKA_RESP_ERR__UNKNOWN_PARTITION if \p partition was not found
* in the list.
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_topic_partition_list_set_offset (
rd_kafka_topic_partition_list_t *rktparlist,
const char *topic, int32_t partition, int64_t offset);
/**
* @brief Find element by \p topic and \p partition.
*
* @returns a pointer to the first matching element, or NULL if not found.
*/
RD_EXPORT
rd_kafka_topic_partition_t *
rd_kafka_topic_partition_list_find (rd_kafka_topic_partition_list_t *rktparlist,
const char *topic, int32_t partition);
/**
* @brief Sort list using comparator \p cmp.
*
* If \p cmp is NULL the default comparator will be used that
* sorts by ascending topic name and partition.
*
*/
RD_EXPORT void
rd_kafka_topic_partition_list_sort (rd_kafka_topic_partition_list_t *rktparlist,
int (*cmp) (const void *a, const void *b,
void *opaque),
void *opaque);
/**@}*/
/**
* @name Var-arg tag types
* @{
*
*/
/**
* @enum rd_kafka_vtype_t
*
* @brief Var-arg tag types
*
* @sa rd_kafka_producev()
*/
typedef enum rd_kafka_vtype_t {
RD_KAFKA_VTYPE_END, /**< va-arg sentinel */
RD_KAFKA_VTYPE_TOPIC, /**< (const char *) Topic name */
RD_KAFKA_VTYPE_RKT, /**< (rd_kafka_topic_t *) Topic handle */
RD_KAFKA_VTYPE_PARTITION, /**< (int32_t) Partition */
RD_KAFKA_VTYPE_VALUE, /**< (void *, size_t) Message value (payload)*/
RD_KAFKA_VTYPE_KEY, /**< (void *, size_t) Message key */
RD_KAFKA_VTYPE_OPAQUE, /**< (void *) Application opaque */
RD_KAFKA_VTYPE_MSGFLAGS, /**< (int) RD_KAFKA_MSG_F_.. flags */
RD_KAFKA_VTYPE_TIMESTAMP, /**< (int64_t) Milliseconds since epoch UTC */
RD_KAFKA_VTYPE_HEADER, /**< (const char *, const void *, ssize_t)
* Message Header */
RD_KAFKA_VTYPE_HEADERS, /**< (rd_kafka_headers_t *) Headers list */
} rd_kafka_vtype_t;
/**
* @brief Convenience macros for rd_kafka_vtype_t that takes the
* correct arguments for each vtype.
*/
/*!
* va-arg end sentinel used to terminate the variable argument list
*/
#define RD_KAFKA_V_END RD_KAFKA_VTYPE_END
/*!
* Topic name (const char *)
*/
#define RD_KAFKA_V_TOPIC(topic) \
_LRK_TYPECHECK(RD_KAFKA_VTYPE_TOPIC, const char *, topic), \
(const char *)topic
/*!
* Topic object (rd_kafka_topic_t *)
*/
#define RD_KAFKA_V_RKT(rkt) \
_LRK_TYPECHECK(RD_KAFKA_VTYPE_RKT, rd_kafka_topic_t *, rkt), \
(rd_kafka_topic_t *)rkt
/*!
* Partition (int32_t)
*/
#define RD_KAFKA_V_PARTITION(partition) \
_LRK_TYPECHECK(RD_KAFKA_VTYPE_PARTITION, int32_t, partition), \
(int32_t)partition
/*!
* Message value/payload pointer and length (void *, size_t)
*/
#define RD_KAFKA_V_VALUE(VALUE,LEN) \
_LRK_TYPECHECK2(RD_KAFKA_VTYPE_VALUE, void *, VALUE, size_t, LEN), \
(void *)VALUE, (size_t)LEN
/*!
* Message key pointer and length (const void *, size_t)
*/
#define RD_KAFKA_V_KEY(KEY,LEN) \
_LRK_TYPECHECK2(RD_KAFKA_VTYPE_KEY, const void *, KEY, size_t, LEN), \
(void *)KEY, (size_t)LEN
/*!
* Message opaque pointer (void *)
* Same as \c produce(.., msg_opaque), and \c rkmessage->_private .
*/
#define RD_KAFKA_V_OPAQUE(opaque) \
_LRK_TYPECHECK(RD_KAFKA_VTYPE_OPAQUE, void *, opaque), \
(void *)opaque
/*!
* Message flags (int)
* @sa RD_KAFKA_MSG_F_COPY, et.al.
*/
#define RD_KAFKA_V_MSGFLAGS(msgflags) \
_LRK_TYPECHECK(RD_KAFKA_VTYPE_MSGFLAGS, int, msgflags), \
(int)msgflags
/*!
* Timestamp in milliseconds since epoch UTC (int64_t).
* A value of 0 will use the current wall-clock time.
*/
#define RD_KAFKA_V_TIMESTAMP(timestamp) \
_LRK_TYPECHECK(RD_KAFKA_VTYPE_TIMESTAMP, int64_t, timestamp), \
(int64_t)timestamp
/*!
* Add Message Header (const char *NAME, const void *VALUE, ssize_t LEN).
* @sa rd_kafka_header_add()
* @remark RD_KAFKA_V_HEADER() and RD_KAFKA_V_HEADERS() MUST NOT be mixed
* in the same call to producev().
*/
#define RD_KAFKA_V_HEADER(NAME,VALUE,LEN) \
_LRK_TYPECHECK3(RD_KAFKA_VTYPE_HEADER, const char *, NAME, \
const void *, VALUE, ssize_t, LEN), \
(const char *)NAME, (const void *)VALUE, (ssize_t)LEN
/*!
* Message Headers list (rd_kafka_headers_t *).
* The message object will assume ownership of the headers (unless producev()
* fails).
* Any existing headers will be replaced.
* @sa rd_kafka_message_set_headers()
* @remark RD_KAFKA_V_HEADER() and RD_KAFKA_V_HEADERS() MUST NOT be mixed
* in the same call to producev().
*/
#define RD_KAFKA_V_HEADERS(HDRS) \
_LRK_TYPECHECK(RD_KAFKA_VTYPE_HEADERS, rd_kafka_headers_t *, HDRS), \
(rd_kafka_headers_t *)HDRS
/**@}*/
/**
* @name Message headers
* @{
*
* @brief Message headers consist of a list of (string key, binary value) pairs.
* Duplicate keys are supported and the order in which keys were
* added are retained.
*
* Header values are considered binary and may have three types of
* value:
* - proper value with size > 0 and a valid pointer
* - empty value with size = 0 and any non-NULL pointer
* - null value with size = 0 and a NULL pointer
*
* Headers require Apache Kafka broker version v0.11.0.0 or later.
*
* Header operations are O(n).
*/
typedef struct rd_kafka_headers_s rd_kafka_headers_t;
/**
* @brief Create a new headers list.
*
* @param initial_count Preallocate space for this number of headers.
* Any number of headers may be added, updated and
* removed regardless of the initial count.
*/
RD_EXPORT rd_kafka_headers_t *rd_kafka_headers_new (size_t initial_count);
/**
* @brief Destroy the headers list. The object and any returned value pointers
* are not usable after this call.
*/
RD_EXPORT void rd_kafka_headers_destroy (rd_kafka_headers_t *hdrs);
/**
* @brief Make a copy of headers list \p src.
*/
RD_EXPORT rd_kafka_headers_t *
rd_kafka_headers_copy (const rd_kafka_headers_t *src);
/**
* @brief Add header with name \p name and value \p val (copied) of size
* \p size (not including null-terminator).
*
* @param name Header name.
* @param name_size Header name size (not including the null-terminator).
* If -1 the \p name length is automatically acquired using
* strlen().
* @param value Pointer to header value, or NULL (set size to 0 or -1).
* @param value_size Size of header value. If -1 the \p value is assumed to be a
* null-terminated string and the length is automatically
* acquired using strlen().
*
* @returns RD_KAFKA_RESP_ERR__READ_ONLY if the headers are read-only,
* else RD_KAFKA_RESP_ERR_NO_ERROR.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_header_add (rd_kafka_headers_t *hdrs,
const char *name, ssize_t name_size,
const void *value, ssize_t value_size);
/**
* @brief Remove all headers for the given key (if any).
*
* @returns RD_KAFKA_RESP_ERR__READ_ONLY if the headers are read-only,
* RD_KAFKA_RESP_ERR__NOENT if no matching headers were found,
* else RD_KAFKA_RESP_ERR_NO_ERROR if headers were removed.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_header_remove (rd_kafka_headers_t *hdrs, const char *name);
/**
* @brief Find last header in list \p hdrs matching \p name.
*
* @param name Header to find (last match).
* @param valuep (out) Set to a (null-terminated) const pointer to the value
* (may be NULL).
* @param sizep (out) Set to the value's size (not including null-terminator).
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR if an entry was found, else
* RD_KAFKA_RESP_ERR__NOENT.
*
* @remark The returned pointer in \p valuep includes a trailing null-terminator
* that is not accounted for in \p sizep.
* @remark The returned pointer is only valid as long as the headers list and
* the header item is valid.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_header_get_last (const rd_kafka_headers_t *hdrs,
const char *name, const void **valuep, size_t *sizep);
/**
* @brief Iterator for headers matching \p name.
*
* Same semantics as rd_kafka_header_get_last()
*
* @param hdrs Headers to iterate.
* @param idx Iterator index, start at 0 and increment by one for each call
* as long as RD_KAFKA_RESP_ERR_NO_ERROR is returned.
* @param name Header name to match.
* @param valuep (out) Set to a (null-terminated) const pointer to the value
* (may be NULL).
* @param sizep (out) Set to the value's size (not including null-terminator).
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_header_get (const rd_kafka_headers_t *hdrs, size_t idx,
const char *name, const void **valuep, size_t *sizep);
/**
* @brief Iterator for all headers.
*
* Same semantics as rd_kafka_header_get()
*
* @sa rd_kafka_header_get()
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_header_get_all (const rd_kafka_headers_t *hdrs, size_t idx,
const char **namep,
const void **valuep, size_t *sizep);
/**@}*/
/**
* @name Kafka messages
* @{
*
*/
// FIXME: This doesn't show up in docs for some reason
// "Compound rd_kafka_message_t is not documented."
/**
* @brief A Kafka message as returned by the \c rd_kafka_consume*() family
* of functions as well as provided to the Producer \c dr_msg_cb().
*
* For the consumer this object has two purposes:
* - provide the application with a consumed message. (\c err == 0)
* - report per-topic+partition consumer errors (\c err != 0)
*
* The application must check \c err to decide what action to take.
*
* When the application is finished with a message it must call
* rd_kafka_message_destroy() unless otherwise noted.
*/
typedef struct rd_kafka_message_s {
rd_kafka_resp_err_t err; /**< Non-zero for error signaling. */
rd_kafka_topic_t *rkt; /**< Topic */
int32_t partition; /**< Partition */
void *payload; /**< Producer: original message payload.
* Consumer: Depends on the value of \c err :
* - \c err==0: Message payload.
* - \c err!=0: Error string */
size_t len; /**< Depends on the value of \c err :
* - \c err==0: Message payload length
* - \c err!=0: Error string length */
void *key; /**< Depends on the value of \c err :
* - \c err==0: Optional message key */
size_t key_len; /**< Depends on the value of \c err :
* - \c err==0: Optional message key length*/
int64_t offset; /**< Consumer:
* - Message offset (or offset for error
* if \c err!=0 if applicable).
* Producer, dr_msg_cb:
* Message offset assigned by broker.
* May be RD_KAFKA_OFFSET_INVALID
* for retried messages when
* idempotence is enabled. */
void *_private; /**< Consume:
* - rdkafka private pointer: DO NOT MODIFY
* - dr_msg_cb:
* msg_opaque from produce() call */
} rd_kafka_message_t;
/**
* @brief Frees resources for \p rkmessage and hands ownership back to rdkafka.
*/
RD_EXPORT
void rd_kafka_message_destroy(rd_kafka_message_t *rkmessage);
/**
* @brief Returns the error string for an errored rd_kafka_message_t or NULL if
* there was no error.
*
* @remark This function MUST NOT be used with the producer.
*/
static RD_INLINE const char *
RD_UNUSED
rd_kafka_message_errstr(const rd_kafka_message_t *rkmessage) {
if (!rkmessage->err)
return NULL;
if (rkmessage->payload)
return (const char *)rkmessage->payload;
return rd_kafka_err2str(rkmessage->err);
}
/**
* @brief Returns the message timestamp for a consumed message.
*
* The timestamp is the number of milliseconds since the epoch (UTC).
*
* \p tstype (if not NULL) is updated to indicate the type of timestamp.
*
* @returns message timestamp, or -1 if not available.
*
* @remark Message timestamps require broker version 0.10.0 or later.
*/
RD_EXPORT
int64_t rd_kafka_message_timestamp (const rd_kafka_message_t *rkmessage,
rd_kafka_timestamp_type_t *tstype);
/**
* @brief Returns the latency for a produced message measured from
* the produce() call.
*
* @returns the latency in microseconds, or -1 if not available.
*/
RD_EXPORT
int64_t rd_kafka_message_latency (const rd_kafka_message_t *rkmessage);
/**
* @brief Get the message header list.
*
* The returned pointer in \p *hdrsp is associated with the \p rkmessage and
* must not be used after destruction of the message object or the header
* list is replaced with rd_kafka_message_set_headers().
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR if headers were returned,
* RD_KAFKA_RESP_ERR__NOENT if the message has no headers,
* or another error code if the headers could not be parsed.
*
* @remark Headers require broker version 0.11.0.0 or later.
*
* @remark As an optimization the raw protocol headers are parsed on
* the first call to this function.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_message_headers (const rd_kafka_message_t *rkmessage,
rd_kafka_headers_t **hdrsp);
/**
* @brief Get the message header list and detach the list from the message
* making the application the owner of the headers.
* The application must eventually destroy the headers using
* rd_kafka_headers_destroy().
* The message's headers will be set to NULL.
*
* Otherwise same semantics as rd_kafka_message_headers()
*
* @sa rd_kafka_message_headers
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_message_detach_headers (rd_kafka_message_t *rkmessage,
rd_kafka_headers_t **hdrsp);
/**
* @brief Replace the message's current headers with a new list.
*
* @param hdrs New header list. The message object assumes ownership of
* the list, the list will be destroyed automatically with
* the message object.
* The new headers list may be updated until the message object
* is passed or returned to librdkafka.
*
* @remark The existing headers object, if any, will be destroyed.
*/
RD_EXPORT
void rd_kafka_message_set_headers (rd_kafka_message_t *rkmessage,
rd_kafka_headers_t *hdrs);
/**
* @brief Returns the number of header key/value pairs
*
* @param hdrs Headers to count
*/
RD_EXPORT size_t rd_kafka_header_cnt (const rd_kafka_headers_t *hdrs);
/**
* @enum rd_kafka_msg_status_t
* @brief Message persistence status can be used by the application to
* find out if a produced message was persisted in the topic log.
*/
typedef enum {
/**< Message was never transmitted to the broker, or failed with
* an error indicating it was not written to the log.
* Application retry risks ordering, but not duplication. */
RD_KAFKA_MSG_STATUS_NOT_PERSISTED = 0,
/**< Message was transmitted to broker, but no acknowledgement was
* received.
* Application retry risks ordering and duplication. */
RD_KAFKA_MSG_STATUS_POSSIBLY_PERSISTED = 1,
/**< Message was written to the log and acknowledged by the broker. */
RD_KAFKA_MSG_STATUS_PERSISTED = 2
} rd_kafka_msg_status_t;
/**
* @brief Returns the message's persistence status in the topic log.
*
* @remark The message status is not available in on_acknowledgement
* interceptors.
*/
RD_EXPORT rd_kafka_msg_status_t
rd_kafka_message_status (const rd_kafka_message_t *rkmessage);
/**@}*/
/**
* @name Configuration interface
* @{
*
* @brief Main/global configuration property interface
*
*/
/**
* @enum rd_kafka_conf_res_t
* @brief Configuration result type
*/
typedef enum {
RD_KAFKA_CONF_UNKNOWN = -2, /**< Unknown configuration name. */
RD_KAFKA_CONF_INVALID = -1, /**< Invalid configuration value. */
RD_KAFKA_CONF_OK = 0 /**< Configuration okay */
} rd_kafka_conf_res_t;
/**
* @brief Create configuration object.
*
* When providing your own configuration to the \c rd_kafka_*_new_*() calls
* the rd_kafka_conf_t objects needs to be created with this function
* which will set up the defaults.
* I.e.:
* @code
* rd_kafka_conf_t *myconf;
* rd_kafka_conf_res_t res;
*
* myconf = rd_kafka_conf_new();
* res = rd_kafka_conf_set(myconf, "socket.timeout.ms", "600",
* errstr, sizeof(errstr));
* if (res != RD_KAFKA_CONF_OK)
* die("%s\n", errstr);
*
* rk = rd_kafka_new(..., myconf);
* @endcode
*
* Please see CONFIGURATION.md for the default settings or use
* rd_kafka_conf_properties_show() to provide the information at runtime.
*
* The properties are identical to the Apache Kafka configuration properties
* whenever possible.
*
* @returns A new rd_kafka_conf_t object with defaults set.
*
* @sa rd_kafka_conf_set(), rd_kafka_conf_destroy()
*/
RD_EXPORT
rd_kafka_conf_t *rd_kafka_conf_new(void);
/**
* @brief Destroys a conf object.
*/
RD_EXPORT
void rd_kafka_conf_destroy(rd_kafka_conf_t *conf);
/**
* @brief Creates a copy/duplicate of configuration object \p conf
*
* @remark Interceptors are NOT copied to the new configuration object.
* @sa rd_kafka_interceptor_f_on_conf_dup
*/
RD_EXPORT
rd_kafka_conf_t *rd_kafka_conf_dup(const rd_kafka_conf_t *conf);
/**
* @brief Same as rd_kafka_conf_dup() but with an array of property name
* prefixes to filter out (ignore) when copying.
*/
RD_EXPORT
rd_kafka_conf_t *rd_kafka_conf_dup_filter (const rd_kafka_conf_t *conf,
size_t filter_cnt,
const char **filter);
/**
* @returns the configuration object used by an rd_kafka_t instance.
* For use with rd_kafka_conf_get(), et.al., to extract configuration
* properties from a running client.
*
* @remark the returned object is read-only and its lifetime is the same
* as the rd_kafka_t object.
*/
RD_EXPORT
const rd_kafka_conf_t *rd_kafka_conf (rd_kafka_t *rk);
/**
* @brief Sets a configuration property.
*
* \p conf must have been previously created with rd_kafka_conf_new().
*
* Fallthrough:
* Topic-level configuration properties may be set using this interface
* in which case they are applied on the \c default_topic_conf.
* If no \c default_topic_conf has been set one will be created.
* Any sub-sequent rd_kafka_conf_set_default_topic_conf() calls will
* replace the current default topic configuration.
*
* @returns \c rd_kafka_conf_res_t to indicate success or failure.
* In case of failure \p errstr is updated to contain a human readable
* error string.
*/
RD_EXPORT
rd_kafka_conf_res_t rd_kafka_conf_set(rd_kafka_conf_t *conf,
const char *name,
const char *value,
char *errstr, size_t errstr_size);
/**
* @brief Enable event sourcing.
* \p events is a bitmask of \c RD_KAFKA_EVENT_* of events to enable
* for consumption by `rd_kafka_queue_poll()`.
*/
RD_EXPORT
void rd_kafka_conf_set_events(rd_kafka_conf_t *conf, int events);
/**
* @brief Generic event callback to be used with the event API to trigger
* callbacks for \c rd_kafka_event_t objects from a background
* thread serving the background queue.
*
* How to use:
* 1. First set the event callback on the configuration object with this
* function, followed by creating an rd_kafka_t instance
* with rd_kafka_new().
* 2. Get the instance's background queue with rd_kafka_queue_get_background()
* and pass it as the reply/response queue to an API that takes an
* event queue, such as rd_kafka_CreateTopics().
* 3. As the response event is ready and enqueued on the background queue the
* event callback will be triggered from the background thread.
* 4. Prior to destroying the client instance, loose your reference to the
* background queue by calling rd_kafka_queue_destroy().
*
* The application must destroy the \c rkev passed to \p event cb using
* rd_kafka_event_destroy().
*
* The \p event_cb \c opaque argument is the opaque set with
* rd_kafka_conf_set_opaque().
*
* @remark This callback is a specialized alternative to the poll-based
* event API described in the Event interface section.
*
* @remark The \p event_cb will be called spontaneously from a background
* thread completely managed by librdkafka.
* Take care to perform proper locking of application objects.
*
* @warning The application MUST NOT call rd_kafka_destroy() from the
* event callback.
*
* @sa rd_kafka_queue_get_background
*/
RD_EXPORT void
rd_kafka_conf_set_background_event_cb (rd_kafka_conf_t *conf,
void (*event_cb) (rd_kafka_t *rk,
rd_kafka_event_t *rkev,
void *opaque));
/**
* @deprecated See rd_kafka_conf_set_dr_msg_cb()
*/
RD_EXPORT
void rd_kafka_conf_set_dr_cb(rd_kafka_conf_t *conf,
void (*dr_cb) (rd_kafka_t *rk,
void *payload, size_t len,
rd_kafka_resp_err_t err,
void *opaque, void *msg_opaque));
/**
* @brief \b Producer: Set delivery report callback in provided \p conf object.
*
* The delivery report callback will be called once for each message
* accepted by rd_kafka_produce() (et.al) with \p err set to indicate
* the result of the produce request.
*
* The callback is called when a message is succesfully produced or
* if librdkafka encountered a permanent failure.
* Delivery errors occur when the retry count is exceeded, when the
* message.timeout.ms timeout is exceeded or there is a permanent error
* like RD_KAFKA_RESP_ERR_UNKNOWN_TOPIC_OR_PART.
*
* An application must call rd_kafka_poll() at regular intervals to
* serve queued delivery report callbacks.
*
* The broker-assigned offset can be retrieved with \c rkmessage->offset
* and the timestamp can be retrieved using rd_kafka_message_timestamp().
*
* @remark The Idempotent Producer may return invalid timestamp
* (RD_KAFKA_TIMESTAMP_NOT_AVAILABLE), and
* and offset (RD_KAFKA_OFFSET_INVALID) for retried messages
* that were previously successfully delivered but not properly
* acknowledged.
*/
RD_EXPORT
void rd_kafka_conf_set_dr_msg_cb(rd_kafka_conf_t *conf,
void (*dr_msg_cb) (rd_kafka_t *rk,
const rd_kafka_message_t *
rkmessage,
void *opaque));
/**
* @brief \b Consumer: Set consume callback for use with rd_kafka_consumer_poll()
*
*/
RD_EXPORT
void rd_kafka_conf_set_consume_cb (rd_kafka_conf_t *conf,
void (*consume_cb) (rd_kafka_message_t *
rkmessage,
void *opaque));
/**
* @brief \b Consumer: Set rebalance callback for use with
* coordinated consumer group balancing.
*
* The \p err field is set to either RD_KAFKA_RESP_ERR__ASSIGN_PARTITIONS
* or RD_KAFKA_RESP_ERR__REVOKE_PARTITIONS and 'partitions'
* contains the full partition set that was either assigned or revoked.
*
* Registering a \p rebalance_cb turns off librdkafka's automatic
* partition assignment/revocation and instead delegates that responsibility
* to the application's \p rebalance_cb.
*
* The rebalance callback is responsible for updating librdkafka's
* assignment set based on the two events: RD_KAFKA_RESP_ERR__ASSIGN_PARTITIONS
* and RD_KAFKA_RESP_ERR__REVOKE_PARTITIONS but should also be able to handle
* arbitrary rebalancing failures where \p err is neither of those.
* @remark In this latter case (arbitrary error), the application must
* call rd_kafka_assign(rk, NULL) to synchronize state.
*
* Without a rebalance callback this is done automatically by librdkafka
* but registering a rebalance callback gives the application flexibility
* in performing other operations along with the assigning/revocation,
* such as fetching offsets from an alternate location (on assign)
* or manually committing offsets (on revoke).
*
* @remark The \p partitions list is destroyed by librdkafka on return
* return from the rebalance_cb and must not be freed or
* saved by the application.
*
* @remark Be careful when modifying the \p partitions list.
* Changing this list should only be done to change the initial
* offsets for each partition.
* But a function like `rd_kafka_position()` might have unexpected
* effects for instance when a consumer gets assigned a partition
* it used to consume at an earlier rebalance. In this case, the
* list of partitions will be updated with the old offset for that
* partition. In this case, it is generally better to pass a copy
* of the list (see `rd_kafka_topic_partition_list_copy()`).
* The result of `rd_kafka_position()` is typically outdated in
* RD_KAFKA_RESP_ERR__ASSIGN_PARTITIONS.
*
* The following example shows the application's responsibilities:
* @code
* static void rebalance_cb (rd_kafka_t *rk, rd_kafka_resp_err_t err,
* rd_kafka_topic_partition_list_t *partitions,
* void *opaque) {
*
* switch (err)
* {
* case RD_KAFKA_RESP_ERR__ASSIGN_PARTITIONS:
* // application may load offets from arbitrary external
* // storage here and update \p partitions
*
* rd_kafka_assign(rk, partitions);
* break;
*
* case RD_KAFKA_RESP_ERR__REVOKE_PARTITIONS:
* if (manual_commits) // Optional explicit manual commit
* rd_kafka_commit(rk, partitions, 0); // sync commit
*
* rd_kafka_assign(rk, NULL);
* break;
*
* default:
* handle_unlikely_error(err);
* rd_kafka_assign(rk, NULL); // sync state
* break;
* }
* }
* @endcode
*/
RD_EXPORT
void rd_kafka_conf_set_rebalance_cb (
rd_kafka_conf_t *conf,
void (*rebalance_cb) (rd_kafka_t *rk,
rd_kafka_resp_err_t err,
rd_kafka_topic_partition_list_t *partitions,
void *opaque));
/**
* @brief \b Consumer: Set offset commit callback for use with consumer groups.
*
* The results of automatic or manual offset commits will be scheduled
* for this callback and is served by rd_kafka_consumer_poll().
*
* If no partitions had valid offsets to commit this callback will be called
* with \p err == RD_KAFKA_RESP_ERR__NO_OFFSET which is not to be considered
* an error.
*
* The \p offsets list contains per-partition information:
* - \c offset: committed offset (attempted)
* - \c err: commit error
*/
RD_EXPORT
void rd_kafka_conf_set_offset_commit_cb (
rd_kafka_conf_t *conf,
void (*offset_commit_cb) (rd_kafka_t *rk,
rd_kafka_resp_err_t err,
rd_kafka_topic_partition_list_t *offsets,
void *opaque));
/**
* @brief Set error callback in provided conf object.
*
* The error callback is used by librdkafka to signal warnings and errors
* back to the application.
*
* These errors should generally be considered informational and non-permanent,
* the client will try to recover automatically from all type of errors.
* Given that the client and cluster configuration is correct the
* application should treat these as temporary errors.
*
* \p error_cb will be triggered with \c err set to RD_KAFKA_RESP_ERR__FATAL
* if a fatal error has been raised; in this case use rd_kafka_fatal_error() to
* retrieve the fatal error code and error string, and then begin terminating
* the client instance.
*
* If no \p error_cb is registered, or RD_KAFKA_EVENT_ERROR has not been set
* with rd_kafka_conf_set_events, then the errors will be logged instead.
*/
RD_EXPORT
void rd_kafka_conf_set_error_cb(rd_kafka_conf_t *conf,
void (*error_cb) (rd_kafka_t *rk, int err,
const char *reason,
void *opaque));
/**
* @brief Set throttle callback.
*
* The throttle callback is used to forward broker throttle times to the
* application for Produce and Fetch (consume) requests.
*
* Callbacks are triggered whenever a non-zero throttle time is returned by
* the broker, or when the throttle time drops back to zero.
*
* An application must call rd_kafka_poll() or rd_kafka_consumer_poll() at
* regular intervals to serve queued callbacks.
*
* @remark Requires broker version 0.9.0 or later.
*/
RD_EXPORT
void rd_kafka_conf_set_throttle_cb (rd_kafka_conf_t *conf,
void (*throttle_cb) (
rd_kafka_t *rk,
const char *broker_name,
int32_t broker_id,
int throttle_time_ms,
void *opaque));
/**
* @brief Set logger callback.
*
* The default is to print to stderr, but a syslog logger is also available,
* see rd_kafka_log_print and rd_kafka_log_syslog for the builtin alternatives.
* Alternatively the application may provide its own logger callback.
* Or pass \p func as NULL to disable logging.
*
* This is the configuration alternative to the deprecated rd_kafka_set_logger()
*
* @remark The log_cb will be called spontaneously from librdkafka's internal
* threads unless logs have been forwarded to a poll queue through
* \c rd_kafka_set_log_queue().
* An application MUST NOT call any librdkafka APIs or do any prolonged
* work in a non-forwarded \c log_cb.
*/
RD_EXPORT
void rd_kafka_conf_set_log_cb(rd_kafka_conf_t *conf,
void (*log_cb) (const rd_kafka_t *rk, int level,
const char *fac, const char *buf));
/**
* @brief Set statistics callback in provided conf object.
*
* The statistics callback is triggered from rd_kafka_poll() every
* \c statistics.interval.ms (needs to be configured separately).
* Function arguments:
* - \p rk - Kafka handle
* - \p json - String containing the statistics data in JSON format
* - \p json_len - Length of \p json string.
* - \p opaque - Application-provided opaque.
*
* For more information on the format of \p json, see
* https://github.com/edenhill/librdkafka/wiki/Statistics
*
* If the application wishes to hold on to the \p json pointer and free
* it at a later time it must return 1 from the \p stats_cb.
* If the application returns 0 from the \p stats_cb then librdkafka
* will immediately free the \p json pointer.
*
* See STATISTICS.md for a full definition of the JSON object.
*/
RD_EXPORT
void rd_kafka_conf_set_stats_cb(rd_kafka_conf_t *conf,
int (*stats_cb) (rd_kafka_t *rk,
char *json,
size_t json_len,
void *opaque));
/**
* @brief Set SASL/OAUTHBEARER token refresh callback in provided conf object.
*
* @param conf the configuration to mutate.
* @param oauthbearer_token_refresh_cb the callback to set; callback function
* arguments:<br>
* \p rk - Kafka handle<br>
* \p oauthbearer_config - Value of configuration property
* sasl.oauthbearer.config.
* \p opaque - Application-provided opaque set via
* rd_kafka_conf_set_opaque()
*
* The SASL/OAUTHBEARER token refresh callback is triggered via rd_kafka_poll()
* whenever OAUTHBEARER is the SASL mechanism and a token needs to be retrieved,
* typically based on the configuration defined in \c sasl.oauthbearer.config.
*
* The callback should invoke rd_kafka_oauthbearer_set_token()
* or rd_kafka_oauthbearer_set_token_failure() to indicate success
* or failure, respectively.
*
* The refresh operation is eventable and may be received via
* rd_kafka_queue_poll() with an event type of
* \c RD_KAFKA_EVENT_OAUTHBEARER_TOKEN_REFRESH.
*
* Note that before any SASL/OAUTHBEARER broker connection can succeed the
* application must call rd_kafka_oauthbearer_set_token() once -- either
* directly or, more typically, by invoking either rd_kafka_poll() or
* rd_kafka_queue_poll() -- in order to cause retrieval of an initial token to
* occur.
*
* An unsecured JWT refresh handler is provided by librdkafka for development
* and testing purposes, it is enabled by setting
* the \c enable.sasl.oauthbearer.unsecure.jwt property to true and is
* mutually exclusive to using a refresh callback.
*/
RD_EXPORT
void rd_kafka_conf_set_oauthbearer_token_refresh_cb (
rd_kafka_conf_t *conf,
void (*oauthbearer_token_refresh_cb) (rd_kafka_t *rk,
const char *oauthbearer_config,
void *opaque));
/**
* @brief Set socket callback.
*
* The socket callback is responsible for opening a socket
* according to the supplied \p domain, \p type and \p protocol.
* The socket shall be created with \c CLOEXEC set in a racefree fashion, if
* possible.
*
* Default:
* - on linux: racefree CLOEXEC
* - others : non-racefree CLOEXEC
*
* @remark The callback will be called from an internal librdkafka thread.
*/
RD_EXPORT
void rd_kafka_conf_set_socket_cb(rd_kafka_conf_t *conf,
int (*socket_cb) (int domain, int type,
int protocol,
void *opaque));
/**
* @brief Set connect callback.
*
* The connect callback is responsible for connecting socket \p sockfd
* to peer address \p addr.
* The \p id field contains the broker identifier.
*
* \p connect_cb shall return 0 on success (socket connected) or an error
* number (errno) on error.
*
* @remark The callback will be called from an internal librdkafka thread.
*/
RD_EXPORT void
rd_kafka_conf_set_connect_cb (rd_kafka_conf_t *conf,
int (*connect_cb) (int sockfd,
const struct sockaddr *addr,
int addrlen,
const char *id,
void *opaque));
/**
* @brief Set close socket callback.
*
* Close a socket (optionally opened with socket_cb()).
*
* @remark The callback will be called from an internal librdkafka thread.
*/
RD_EXPORT void
rd_kafka_conf_set_closesocket_cb (rd_kafka_conf_t *conf,
int (*closesocket_cb) (int sockfd,
void *opaque));
#ifndef _MSC_VER
/**
* @brief Set open callback.
*
* The open callback is responsible for opening the file specified by
* pathname, flags and mode.
* The file shall be opened with \c CLOEXEC set in a racefree fashion, if
* possible.
*
* Default:
* - on linux: racefree CLOEXEC
* - others : non-racefree CLOEXEC
*
* @remark The callback will be called from an internal librdkafka thread.
*/
RD_EXPORT
void rd_kafka_conf_set_open_cb (rd_kafka_conf_t *conf,
int (*open_cb) (const char *pathname,
int flags, mode_t mode,
void *opaque));
#endif
/**
* @brief Sets the verification callback of the broker certificate
*
* The verification callback is triggered from internal librdkafka threads
* upon connecting to a broker. On each connection attempt the callback
* will be called for each certificate in the broker's certificate chain,
* starting at the root certification, as long as the application callback
* returns 1 (valid certificate).
* \c broker_name and \c broker_id correspond to the broker the connection
* is being made to.
* The \c x509_error argument indicates if OpenSSL's verification of
* the certificate succeed (0) or failed (an OpenSSL error code).
* The application may set the SSL context error code by returning 0
* from the verify callback and providing a non-zero SSL context error code
* in \p x509_error.
* If the verify callback sets \x509_error to 0, returns 1, and the
* original \p x509_error was non-zero, the error on the SSL context will
* be cleared.
* \p x509_error is always a valid pointer to an int.
*
* \c depth is the depth of the current certificate in the chain, starting
* at the root certificate.
*
* The certificate itself is passed in binary DER format in \c buf of
* size \c size.
*
* The callback must return 1 if verification succeeds, or
* 0 if verification fails and then write a human-readable error message
* to \c errstr (limited to \c errstr_size bytes, including nul-term).
*
* @returns RD_KAFKA_CONF_OK if SSL is supported in this build, else
* RD_KAFKA_CONF_INVALID.
*
* @warning This callback will be called from internal librdkafka threads.
*
* @remark See <openssl/x509_vfy.h> in the OpenSSL source distribution
* for a list of \p x509_error codes.
*/
RD_EXPORT
rd_kafka_conf_res_t rd_kafka_conf_set_ssl_cert_verify_cb (
rd_kafka_conf_t *conf,
int (*ssl_cert_verify_cb) (rd_kafka_t *rk,
const char *broker_name,
int32_t broker_id,
int *x509_error,
int depth,
const char *buf, size_t size,
char *errstr, size_t errstr_size,
void *opaque));
/**
* @enum rd_kafka_cert_type_t
*
* @brief SSL certificate type
*
* @sa rd_kafka_conf_set_ssl_cert
*/
typedef enum rd_kafka_cert_type_t {
RD_KAFKA_CERT_PUBLIC_KEY, /**< Client's public key */
RD_KAFKA_CERT_PRIVATE_KEY, /**< Client's private key */
RD_KAFKA_CERT_CA, /**< CA certificate */
RD_KAFKA_CERT__CNT,
} rd_kafka_cert_type_t;
/**
* @enum rd_kafka_cert_enc_t
*
* @brief SSL certificate encoding
*
* @sa rd_kafka_conf_set_ssl_cert
*/
typedef enum rd_kafka_cert_enc_t {
RD_KAFKA_CERT_ENC_PKCS12, /**< PKCS#12 */
RD_KAFKA_CERT_ENC_DER, /**< DER / binary X.509 ASN1 */
RD_KAFKA_CERT_ENC_PEM, /**< PEM */
RD_KAFKA_CERT_ENC__CNT,
} rd_kafka_cert_enc_t;
/**
* @brief Set certificate/key \p cert_type from the \p cert_enc encoded
* memory at \p buffer of \p size bytes.
*
* @param conf Configuration object.
* @param cert_type Certificate or key type to configure.
* @param cert_enc Buffer \p encoding type.
* @param buffer Memory pointer to encoded certificate or key.
* The memory is not referenced after this function returns.
* @param size Size of memory at \p buffer.
* @param errstr Memory were a human-readable error string will be written
* on failure.
* @param errstr_size Size of \p errstr, including space for nul-terminator.
*
* @returns RD_KAFKA_CONF_OK on success or RD_KAFKA_CONF_INVALID if the
* memory in \p buffer is of incorrect encoding, or if librdkafka
* was not built with SSL support.
*
* @remark Calling this method multiple times with the same \p cert_type
* will replace the previous value.
*
* @remark Calling this method with \p buffer set to NULL will clear the
* configuration for \p cert_type.
*
* @remark The private key may require a password, which must be specified
* with the `ssl.key.password` configuration property prior to
* calling this function.
*
* @remark Private and public keys in PEM format may also be set with the
* `ssl.key.pem` and `ssl.certificate.pem` configuration properties.
*/
RD_EXPORT rd_kafka_conf_res_t
rd_kafka_conf_set_ssl_cert (rd_kafka_conf_t *conf,
rd_kafka_cert_type_t cert_type,
rd_kafka_cert_enc_t cert_enc,
const void *buffer, size_t size,
char *errstr, size_t errstr_size);
/**
* @brief Sets the application's opaque pointer that will be passed to callbacks
*/
RD_EXPORT
void rd_kafka_conf_set_opaque(rd_kafka_conf_t *conf, void *opaque);
/**
* @brief Retrieves the opaque pointer previously set with rd_kafka_conf_set_opaque()
*/
RD_EXPORT
void *rd_kafka_opaque(const rd_kafka_t *rk);
/**
* @brief Sets the default topic configuration to use for automatically
* subscribed topics (e.g., through pattern-matched topics).
* The topic config object is not usable after this call.
*
* @warning Any topic configuration settings that have been set on the
* global rd_kafka_conf_t object will be overwritten by this call
* since the implicitly created default topic config object is
* replaced by the user-supplied one.
*
* @deprecated Set default topic level configuration on the
* global rd_kafka_conf_t object instead.
*/
RD_EXPORT
void rd_kafka_conf_set_default_topic_conf (rd_kafka_conf_t *conf,
rd_kafka_topic_conf_t *tconf);
/**
* @brief Retrieve configuration value for property \p name.
*
* If \p dest is non-NULL the value will be written to \p dest with at
* most \p dest_size.
*
* \p *dest_size is updated to the full length of the value, thus if
* \p *dest_size initially is smaller than the full length the application
* may reallocate \p dest to fit the returned \p *dest_size and try again.
*
* If \p dest is NULL only the full length of the value is returned.
*
* Fallthrough:
* Topic-level configuration properties from the \c default_topic_conf
* may be retrieved using this interface.
*
* @returns \p RD_KAFKA_CONF_OK if the property name matched, else
* \p RD_KAFKA_CONF_UNKNOWN.
*/
RD_EXPORT
rd_kafka_conf_res_t rd_kafka_conf_get (const rd_kafka_conf_t *conf,
const char *name,
char *dest, size_t *dest_size);
/**
* @brief Retrieve topic configuration value for property \p name.
*
* @sa rd_kafka_conf_get()
*/
RD_EXPORT
rd_kafka_conf_res_t rd_kafka_topic_conf_get (const rd_kafka_topic_conf_t *conf,
const char *name,
char *dest, size_t *dest_size);
/**
* @brief Dump the configuration properties and values of \p conf to an array
* with \"key\", \"value\" pairs.
*
* The number of entries in the array is returned in \p *cntp.
*
* The dump must be freed with `rd_kafka_conf_dump_free()`.
*/
RD_EXPORT
const char **rd_kafka_conf_dump(rd_kafka_conf_t *conf, size_t *cntp);
/**
* @brief Dump the topic configuration properties and values of \p conf
* to an array with \"key\", \"value\" pairs.
*
* The number of entries in the array is returned in \p *cntp.
*
* The dump must be freed with `rd_kafka_conf_dump_free()`.
*/
RD_EXPORT
const char **rd_kafka_topic_conf_dump(rd_kafka_topic_conf_t *conf,
size_t *cntp);
/**
* @brief Frees a configuration dump returned from `rd_kafka_conf_dump()` or
* `rd_kafka_topic_conf_dump().
*/
RD_EXPORT
void rd_kafka_conf_dump_free(const char **arr, size_t cnt);
/**
* @brief Prints a table to \p fp of all supported configuration properties,
* their default values as well as a description.
*/
RD_EXPORT
void rd_kafka_conf_properties_show(FILE *fp);
/**@}*/
/**
* @name Topic configuration
* @{
*
* @brief Topic configuration property interface
*
*/
/**
* @brief Create topic configuration object
*
* @sa Same semantics as for rd_kafka_conf_new().
*/
RD_EXPORT
rd_kafka_topic_conf_t *rd_kafka_topic_conf_new(void);
/**
* @brief Creates a copy/duplicate of topic configuration object \p conf.
*/
RD_EXPORT
rd_kafka_topic_conf_t *rd_kafka_topic_conf_dup(const rd_kafka_topic_conf_t
*conf);
/**
* @brief Creates a copy/duplicate of \p rk 's default topic configuration
* object.
*/
RD_EXPORT
rd_kafka_topic_conf_t *rd_kafka_default_topic_conf_dup (rd_kafka_t *rk);
/**
* @brief Destroys a topic conf object.
*/
RD_EXPORT
void rd_kafka_topic_conf_destroy(rd_kafka_topic_conf_t *topic_conf);
/**
* @brief Sets a single rd_kafka_topic_conf_t value by property name.
*
* \p topic_conf should have been previously set up
* with `rd_kafka_topic_conf_new()`.
*
* @returns rd_kafka_conf_res_t to indicate success or failure.
*/
RD_EXPORT
rd_kafka_conf_res_t rd_kafka_topic_conf_set(rd_kafka_topic_conf_t *conf,
const char *name,
const char *value,
char *errstr, size_t errstr_size);
/**
* @brief Sets the application's opaque pointer that will be passed to all topic
* callbacks as the \c rkt_opaque argument.
*/
RD_EXPORT
void rd_kafka_topic_conf_set_opaque(rd_kafka_topic_conf_t *conf, void *opaque);
/**
* @brief \b Producer: Set partitioner callback in provided topic conf object.
*
* The partitioner may be called in any thread at any time,
* it may be called multiple times for the same message/key.
*
* Partitioner function constraints:
* - MUST NOT call any rd_kafka_*() functions except:
* rd_kafka_topic_partition_available()
* - MUST NOT block or execute for prolonged periods of time.
* - MUST return a value between 0 and partition_cnt-1, or the
* special \c RD_KAFKA_PARTITION_UA value if partitioning
* could not be performed.
*/
RD_EXPORT
void
rd_kafka_topic_conf_set_partitioner_cb (rd_kafka_topic_conf_t *topic_conf,
int32_t (*partitioner) (
const rd_kafka_topic_t *rkt,
const void *keydata,
size_t keylen,
int32_t partition_cnt,
void *rkt_opaque,
void *msg_opaque));
/**
* @brief \b Producer: Set message queueing order comparator callback.
*
* The callback may be called in any thread at any time,
* it may be called multiple times for the same message.
*
* Ordering comparator function constraints:
* - MUST be stable sort (same input gives same output).
* - MUST NOT call any rd_kafka_*() functions.
* - MUST NOT block or execute for prolonged periods of time.
*
* The comparator shall compare the two messages and return:
* - < 0 if message \p a should be inserted before message \p b.
* - >=0 if message \p a should be inserted after message \p b.
*
* @remark Insert sorting will be used to enqueue the message in the
* correct queue position, this comes at a cost of O(n).
*
* @remark If `queuing.strategy=fifo` new messages are enqueued to the
* tail of the queue regardless of msg_order_cmp, but retried messages
* are still affected by msg_order_cmp.
*
* @warning THIS IS AN EXPERIMENTAL API, SUBJECT TO CHANGE OR REMOVAL,
* DO NOT USE IN PRODUCTION.
*/
RD_EXPORT void
rd_kafka_topic_conf_set_msg_order_cmp (rd_kafka_topic_conf_t *topic_conf,
int (*msg_order_cmp) (
const rd_kafka_message_t *a,
const rd_kafka_message_t *b));
/**
* @brief Check if partition is available (has a leader broker).
*
* @returns 1 if the partition is available, else 0.
*
* @warning This function must only be called from inside a partitioner function
*/
RD_EXPORT
int rd_kafka_topic_partition_available(const rd_kafka_topic_t *rkt,
int32_t partition);
/*******************************************************************
* *
* Partitioners provided by rdkafka *
* *
*******************************************************************/
/**
* @brief Random partitioner.
*
* Will try not to return unavailable partitions.
*
* @returns a random partition between 0 and \p partition_cnt - 1.
*
*/
RD_EXPORT
int32_t rd_kafka_msg_partitioner_random(const rd_kafka_topic_t *rkt,
const void *key, size_t keylen,
int32_t partition_cnt,
void *opaque, void *msg_opaque);
/**
* @brief Consistent partitioner.
*
* Uses consistent hashing to map identical keys onto identical partitions.
*
* @returns a \"random\" partition between 0 and \p partition_cnt - 1 based on
* the CRC value of the key
*/
RD_EXPORT
int32_t rd_kafka_msg_partitioner_consistent (const rd_kafka_topic_t *rkt,
const void *key, size_t keylen,
int32_t partition_cnt,
void *opaque, void *msg_opaque);
/**
* @brief Consistent-Random partitioner.
*
* This is the default partitioner.
* Uses consistent hashing to map identical keys onto identical partitions, and
* messages without keys will be assigned via the random partitioner.
*
* @returns a \"random\" partition between 0 and \p partition_cnt - 1 based on
* the CRC value of the key (if provided)
*/
RD_EXPORT
int32_t rd_kafka_msg_partitioner_consistent_random (const rd_kafka_topic_t *rkt,
const void *key, size_t keylen,
int32_t partition_cnt,
void *opaque, void *msg_opaque);
/**
* @brief Murmur2 partitioner (Java compatible).
*
* Uses consistent hashing to map identical keys onto identical partitions
* using Java-compatible Murmur2 hashing.
*
* @returns a partition between 0 and \p partition_cnt - 1.
*/
RD_EXPORT
int32_t rd_kafka_msg_partitioner_murmur2 (const rd_kafka_topic_t *rkt,
const void *key, size_t keylen,
int32_t partition_cnt,
void *rkt_opaque,
void *msg_opaque);
/**
* @brief Consistent-Random Murmur2 partitioner (Java compatible).
*
* Uses consistent hashing to map identical keys onto identical partitions
* using Java-compatible Murmur2 hashing.
* Messages without keys will be assigned via the random partitioner.
*
* @returns a partition between 0 and \p partition_cnt - 1.
*/
RD_EXPORT
int32_t rd_kafka_msg_partitioner_murmur2_random (const rd_kafka_topic_t *rkt,
const void *key, size_t keylen,
int32_t partition_cnt,
void *rkt_opaque,
void *msg_opaque);
/**@}*/
/**
* @name Main Kafka and Topic object handles
* @{
*
*
*/
/**
* @brief Creates a new Kafka handle and starts its operation according to the
* specified \p type (\p RD_KAFKA_CONSUMER or \p RD_KAFKA_PRODUCER).
*
* \p conf is an optional struct created with `rd_kafka_conf_new()` that will
* be used instead of the default configuration.
* The \p conf object is freed by this function on success and must not be used
* or destroyed by the application sub-sequently.
* See `rd_kafka_conf_set()` et.al for more information.
*
* \p errstr must be a pointer to memory of at least size \p errstr_size where
* `rd_kafka_new()` may write a human readable error message in case the
* creation of a new handle fails. In which case the function returns NULL.
*
* @remark \b RD_KAFKA_CONSUMER: When a new \p RD_KAFKA_CONSUMER
* rd_kafka_t handle is created it may either operate in the
* legacy simple consumer mode using the rd_kafka_consume_start()
* interface, or the High-level KafkaConsumer API.
* @remark An application must only use one of these groups of APIs on a given
* rd_kafka_t RD_KAFKA_CONSUMER handle.
*
* @returns The Kafka handle on success or NULL on error (see \p errstr)
*
* @sa To destroy the Kafka handle, use rd_kafka_destroy().
*/
RD_EXPORT
rd_kafka_t *rd_kafka_new(rd_kafka_type_t type, rd_kafka_conf_t *conf,
char *errstr, size_t errstr_size);
/**
* @brief Destroy Kafka handle.
*
* @remark This is a blocking operation.
* @remark rd_kafka_consumer_close() will be called from this function
* if the instance type is RD_KAFKA_CONSUMER, a \c group.id was
* configured, and the rd_kafka_consumer_close() was not
* explicitly called by the application. This in turn may
* trigger consumer callbacks, such as rebalance_cb.
* Use rd_kafka_destroy_flags() with
* RD_KAFKA_DESTROY_F_NO_CONSUMER_CLOSE to avoid this behaviour.
*
* @sa rd_kafka_destroy_flags()
*/
RD_EXPORT
void rd_kafka_destroy(rd_kafka_t *rk);
/**
* @brief Destroy Kafka handle according to specified destroy flags
*
*/
RD_EXPORT
void rd_kafka_destroy_flags (rd_kafka_t *rk, int flags);
/**
* @brief Flags for rd_kafka_destroy_flags()
*/
/*!
* Don't call consumer_close() to leave group and commit final offsets.
*
* This also disables consumer callbacks to be called from rd_kafka_destroy*(),
* such as rebalance_cb.
*
* The consumer group handler is still closed internally, but from an
* application perspective none of the functionality from consumer_close()
* is performed.
*/
#define RD_KAFKA_DESTROY_F_NO_CONSUMER_CLOSE 0x8
/**
* @brief Returns Kafka handle name.
*/
RD_EXPORT
const char *rd_kafka_name(const rd_kafka_t *rk);
/**
* @brief Returns Kafka handle type.
*/
RD_EXPORT
rd_kafka_type_t rd_kafka_type(const rd_kafka_t *rk);
/**
* @brief Returns this client's broker-assigned group member id
*
* @remark This currently requires the high-level KafkaConsumer
*
* @returns An allocated string containing the current broker-assigned group
* member id, or NULL if not available.
* The application must free the string with \p free() or
* rd_kafka_mem_free()
*/
RD_EXPORT
char *rd_kafka_memberid (const rd_kafka_t *rk);
/**
* @brief Returns the ClusterId as reported in broker metadata.
*
* @param timeout_ms If there is no cached value from metadata retrieval
* then this specifies the maximum amount of time
* (in milliseconds) the call will block waiting
* for metadata to be retrieved.
* Use 0 for non-blocking calls.
* @remark Requires broker version >=0.10.0 and api.version.request=true.
*
* @remark The application must free the returned pointer
* using rd_kafka_mem_free().
*
* @returns a newly allocated string containing the ClusterId, or NULL
* if no ClusterId could be retrieved in the allotted timespan.
*/
RD_EXPORT
char *rd_kafka_clusterid (rd_kafka_t *rk, int timeout_ms);
/**
* @brief Returns the current ControllerId as reported in broker metadata.
*
* @param timeout_ms If there is no cached value from metadata retrieval
* then this specifies the maximum amount of time
* (in milliseconds) the call will block waiting
* for metadata to be retrieved.
* Use 0 for non-blocking calls.
* @remark Requires broker version >=0.10.0 and api.version.request=true.
*
* @returns the controller broker id (>= 0), or -1 if no ControllerId could be
* retrieved in the allotted timespan.
*/
RD_EXPORT
int32_t rd_kafka_controllerid (rd_kafka_t *rk, int timeout_ms);
/**
* @brief Creates a new topic handle for topic named \p topic.
*
* \p conf is an optional configuration for the topic created with
* `rd_kafka_topic_conf_new()` that will be used instead of the default
* topic configuration.
* The \p conf object is freed by this function and must not be used or
* destroyed by the application sub-sequently.
* See `rd_kafka_topic_conf_set()` et.al for more information.
*
* Topic handles are refcounted internally and calling rd_kafka_topic_new()
* again with the same topic name will return the previous topic handle
* without updating the original handle's configuration.
* Applications must eventually call rd_kafka_topic_destroy() for each
* succesfull call to rd_kafka_topic_new() to clear up resources.
*
* @returns the new topic handle or NULL on error (use rd_kafka_errno2err()
* to convert system \p errno to an rd_kafka_resp_err_t error code.
*
* @sa rd_kafka_topic_destroy()
*/
RD_EXPORT
rd_kafka_topic_t *rd_kafka_topic_new(rd_kafka_t *rk, const char *topic,
rd_kafka_topic_conf_t *conf);
/**
* @brief Loose application's topic handle refcount as previously created
* with `rd_kafka_topic_new()`.
*
* @remark Since topic objects are refcounted (both internally and for the app)
* the topic object might not actually be destroyed by this call,
* but the application must consider the object destroyed.
*/
RD_EXPORT
void rd_kafka_topic_destroy(rd_kafka_topic_t *rkt);
/**
* @brief Returns the topic name.
*/
RD_EXPORT
const char *rd_kafka_topic_name(const rd_kafka_topic_t *rkt);
/**
* @brief Get the \p rkt_opaque pointer that was set in the topic configuration.
*/
RD_EXPORT
void *rd_kafka_topic_opaque (const rd_kafka_topic_t *rkt);
/**
* @brief Unassigned partition.
*
* The unassigned partition is used by the producer API for messages
* that should be partitioned using the configured or default partitioner.
*/
#define RD_KAFKA_PARTITION_UA ((int32_t)-1)
/**
* @brief Polls the provided kafka handle for events.
*
* Events will cause application provided callbacks to be called.
*
* The \p timeout_ms argument specifies the maximum amount of time
* (in milliseconds) that the call will block waiting for events.
* For non-blocking calls, provide 0 as \p timeout_ms.
* To wait indefinately for an event, provide -1.
*
* @remark An application should make sure to call poll() at regular
* intervals to serve any queued callbacks waiting to be called.
* @remark If your producer doesn't have any callback set (in particular
* via rd_kafka_conf_set_dr_msg_cb or rd_kafka_conf_set_error_cb)
* you might chose not to call poll(), though this is not
* recommended.
*
* Events:
* - delivery report callbacks (if dr_cb/dr_msg_cb is configured) [producer]
* - error callbacks (rd_kafka_conf_set_error_cb()) [all]
* - stats callbacks (rd_kafka_conf_set_stats_cb()) [all]
* - throttle callbacks (rd_kafka_conf_set_throttle_cb()) [all]
* - OAUTHBEARER token refresh callbacks (rd_kafka_conf_set_oauthbearer_token_refresh_cb()) [all]
*
* @returns the number of events served.
*/
RD_EXPORT
int rd_kafka_poll(rd_kafka_t *rk, int timeout_ms);
/**
* @brief Cancels the current callback dispatcher (rd_kafka_poll(),
* rd_kafka_consume_callback(), etc).
*
* A callback may use this to force an immediate return to the calling
* code (caller of e.g. rd_kafka_poll()) without processing any further
* events.
*
* @remark This function MUST ONLY be called from within a librdkafka callback.
*/
RD_EXPORT
void rd_kafka_yield (rd_kafka_t *rk);
/**
* @brief Pause producing or consumption for the provided list of partitions.
*
* Success or error is returned per-partition \p err in the \p partitions list.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_pause_partitions (rd_kafka_t *rk,
rd_kafka_topic_partition_list_t *partitions);
/**
* @brief Resume producing consumption for the provided list of partitions.
*
* Success or error is returned per-partition \p err in the \p partitions list.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_resume_partitions (rd_kafka_t *rk,
rd_kafka_topic_partition_list_t *partitions);
/**
* @brief Query broker for low (oldest/beginning) and high (newest/end) offsets
* for partition.
*
* Offsets are returned in \p *low and \p *high respectively.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success or an error code on failure.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_query_watermark_offsets (rd_kafka_t *rk,
const char *topic, int32_t partition,
int64_t *low, int64_t *high, int timeout_ms);
/**
* @brief Get last known low (oldest/beginning) and high (newest/end) offsets
* for partition.
*
* The low offset is updated periodically (if statistics.interval.ms is set)
* while the high offset is updated on each fetched message set from the broker.
*
* If there is no cached offset (either low or high, or both) then
* RD_KAFKA_OFFSET_INVALID will be returned for the respective offset.
*
* Offsets are returned in \p *low and \p *high respectively.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success or an error code on failure.
*
* @remark Shall only be used with an active consumer instance.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_get_watermark_offsets (rd_kafka_t *rk,
const char *topic, int32_t partition,
int64_t *low, int64_t *high);
/**
* @brief Look up the offsets for the given partitions by timestamp.
*
* The returned offset for each partition is the earliest offset whose
* timestamp is greater than or equal to the given timestamp in the
* corresponding partition.
*
* The timestamps to query are represented as \c offset in \p offsets
* on input, and \c offset will contain the offset on output.
*
* The function will block for at most \p timeout_ms milliseconds.
*
* @remark Duplicate Topic+Partitions are not supported.
* @remark Per-partition errors may be returned in \c rd_kafka_topic_partition_t.err
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR if offsets were be queried (do note
* that per-partition errors might be set),
* RD_KAFKA_RESP_ERR__TIMED_OUT if not all offsets could be fetched
* within \p timeout_ms,
* RD_KAFKA_RESP_ERR__INVALID_ARG if the \p offsets list is empty,
* RD_KAFKA_RESP_ERR__UNKNOWN_PARTITION if all partitions are unknown,
* RD_KAFKA_RESP_ERR_LEADER_NOT_AVAILABLE if unable to query leaders
* for the given partitions.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_offsets_for_times (rd_kafka_t *rk,
rd_kafka_topic_partition_list_t *offsets,
int timeout_ms);
/**
* @brief Free pointer returned by librdkafka
*
* This is typically an abstraction for the free(3) call and makes sure
* the application can use the same memory allocator as librdkafka for
* freeing pointers returned by librdkafka.
*
* In standard setups it is usually not necessary to use this interface
* rather than the free(3) functione.
*
* \p rk must be set for memory returned by APIs that take an \c rk argument,
* for other APIs pass NULL for \p rk.
*
* @remark rd_kafka_mem_free() must only be used for pointers returned by APIs
* that explicitly mention using this function for freeing.
*/
RD_EXPORT
void rd_kafka_mem_free (rd_kafka_t *rk, void *ptr);
/**@}*/
/**
* @name Queue API
* @{
*
* Message queues allows the application to re-route consumed messages
* from multiple topic+partitions into one single queue point.
* This queue point containing messages from a number of topic+partitions
* may then be served by a single rd_kafka_consume*_queue() call,
* rather than one call per topic+partition combination.
*/
/**
* @brief Create a new message queue.
*
* See rd_kafka_consume_start_queue(), rd_kafka_consume_queue(), et.al.
*/
RD_EXPORT
rd_kafka_queue_t *rd_kafka_queue_new(rd_kafka_t *rk);
/**
* Destroy a queue, purging all of its enqueued messages.
*/
RD_EXPORT
void rd_kafka_queue_destroy(rd_kafka_queue_t *rkqu);
/**
* @returns a reference to the main librdkafka event queue.
* This is the queue served by rd_kafka_poll().
*
* Use rd_kafka_queue_destroy() to loose the reference.
*/
RD_EXPORT
rd_kafka_queue_t *rd_kafka_queue_get_main (rd_kafka_t *rk);
/**
* @returns a reference to the librdkafka consumer queue.
* This is the queue served by rd_kafka_consumer_poll().
*
* Use rd_kafka_queue_destroy() to loose the reference.
*
* @remark rd_kafka_queue_destroy() MUST be called on this queue
* prior to calling rd_kafka_consumer_close().
*/
RD_EXPORT
rd_kafka_queue_t *rd_kafka_queue_get_consumer (rd_kafka_t *rk);
/**
* @returns a reference to the partition's queue, or NULL if
* partition is invalid.
*
* Use rd_kafka_queue_destroy() to loose the reference.
*
* @remark rd_kafka_queue_destroy() MUST be called on this queue
*
* @remark This function only works on consumers.
*/
RD_EXPORT
rd_kafka_queue_t *rd_kafka_queue_get_partition (rd_kafka_t *rk,
const char *topic,
int32_t partition);
/**
* @returns a reference to the background thread queue, or NULL if the
* background queue is not enabled.
*
* To enable the background thread queue set a generic event handler callback
* with rd_kafka_conf_set_background_event_cb() on the client instance
* configuration object (rd_kafka_conf_t).
*
* The background queue is polled and served by librdkafka and MUST NOT be
* polled, forwarded, or otherwise managed by the application, it may only
* be used as the destination queue passed to queue-enabled APIs, such as
* the Admin API.
*
* The background thread queue provides the application with an automatically
* polled queue that triggers the event callback in a background thread,
* this background thread is completely managed by librdkafka.
*
* Use rd_kafka_queue_destroy() to loose the reference.
*
* @warning The background queue MUST NOT be read from (polled, consumed, etc),
* or forwarded from.
*/
RD_EXPORT
rd_kafka_queue_t *rd_kafka_queue_get_background (rd_kafka_t *rk);
/**
* @brief Forward/re-route queue \p src to \p dst.
* If \p dst is \c NULL the forwarding is removed.
*
* The internal refcounts for both queues are increased.
*
* @remark Regardless of whether \p dst is NULL or not, after calling this
* function, \p src will not forward it's fetch queue to the consumer
* queue.
*/
RD_EXPORT
void rd_kafka_queue_forward (rd_kafka_queue_t *src, rd_kafka_queue_t *dst);
/**
* @brief Forward librdkafka logs (and debug) to the specified queue
* for serving with one of the ..poll() calls.
*
* This allows an application to serve log callbacks (\c log_cb)
* in its thread of choice.
*
* @param rkqu Queue to forward logs to. If the value is NULL the logs
* are forwarded to the main queue.
*
* @remark The configuration property \c log.queue MUST also be set to true.
*
* @remark librdkafka maintains its own reference to the provided queue.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success or an error code on error.
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_set_log_queue (rd_kafka_t *rk,
rd_kafka_queue_t *rkqu);
/**
* @returns the current number of elements in queue.
*/
RD_EXPORT
size_t rd_kafka_queue_length (rd_kafka_queue_t *rkqu);
/**
* @brief Enable IO event triggering for queue.
*
* To ease integration with IO based polling loops this API
* allows an application to create a separate file-descriptor
* that librdkafka will write \p payload (of size \p size) to
* whenever a new element is enqueued on a previously empty queue.
*
* To remove event triggering call with \p fd = -1.
*
* librdkafka will maintain a copy of the \p payload.
*
* @remark IO and callback event triggering are mutually exclusive.
* @remark When using forwarded queues the IO event must only be enabled
* on the final forwarded-to (destination) queue.
*/
RD_EXPORT
void rd_kafka_queue_io_event_enable (rd_kafka_queue_t *rkqu, int fd,
const void *payload, size_t size);
/**
* @brief Enable callback event triggering for queue.
*
* The callback will be called from an internal librdkafka thread
* when a new element is enqueued on a previously empty queue.
*
* To remove event triggering call with \p event_cb = NULL.
*
* @remark IO and callback event triggering are mutually exclusive.
* @remark Since the callback may be triggered from internal librdkafka
* threads, the application must not perform any pro-longed work in
* the callback, or call any librdkafka APIs (for the same rd_kafka_t
* handle).
*/
RD_EXPORT
void rd_kafka_queue_cb_event_enable (rd_kafka_queue_t *rkqu,
void (*event_cb) (rd_kafka_t *rk,
void *opaque),
void *opaque);
/**@}*/
/**
*
* @name Simple Consumer API (legacy)
* @{
*
*/
#define RD_KAFKA_OFFSET_BEGINNING -2 /**< Start consuming from beginning of
* kafka partition queue: oldest msg */
#define RD_KAFKA_OFFSET_END -1 /**< Start consuming from end of kafka
* partition queue: next msg */
#define RD_KAFKA_OFFSET_STORED -1000 /**< Start consuming from offset retrieved
* from offset store */
#define RD_KAFKA_OFFSET_INVALID -1001 /**< Invalid offset */
/** @cond NO_DOC */
#define RD_KAFKA_OFFSET_TAIL_BASE -2000 /* internal: do not use */
/** @endcond */
/**
* @brief Start consuming \p CNT messages from topic's current end offset.
*
* That is, if current end offset is 12345 and \p CNT is 200, it will start
* consuming from offset \c 12345-200 = \c 12145. */
#define RD_KAFKA_OFFSET_TAIL(CNT) (RD_KAFKA_OFFSET_TAIL_BASE - (CNT))
/**
* @brief Start consuming messages for topic \p rkt and \p partition
* at offset \p offset which may either be an absolute \c (0..N)
* or one of the logical offsets:
* - RD_KAFKA_OFFSET_BEGINNING
* - RD_KAFKA_OFFSET_END
* - RD_KAFKA_OFFSET_STORED
* - RD_KAFKA_OFFSET_TAIL
*
* rdkafka will attempt to keep \c queued.min.messages (config property)
* messages in the local queue by repeatedly fetching batches of messages
* from the broker until the threshold is reached.
*
* The application shall use one of the `rd_kafka_consume*()` functions
* to consume messages from the local queue, each kafka message being
* represented as a `rd_kafka_message_t *` object.
*
* `rd_kafka_consume_start()` must not be called multiple times for the same
* topic and partition without stopping consumption first with
* `rd_kafka_consume_stop()`.
*
* @returns 0 on success or -1 on error in which case errno is set accordingly:
* - EBUSY - Conflicts with an existing or previous subscription
* (RD_KAFKA_RESP_ERR__CONFLICT)
* - EINVAL - Invalid offset, or incomplete configuration (lacking group.id)
* (RD_KAFKA_RESP_ERR__INVALID_ARG)
* - ESRCH - requested \p partition is invalid.
* (RD_KAFKA_RESP_ERR__UNKNOWN_PARTITION)
* - ENOENT - topic is unknown in the Kafka cluster.
* (RD_KAFKA_RESP_ERR__UNKNOWN_TOPIC)
*
* Use `rd_kafka_errno2err()` to convert sytem \c errno to `rd_kafka_resp_err_t`
*/
RD_EXPORT
int rd_kafka_consume_start(rd_kafka_topic_t *rkt, int32_t partition,
int64_t offset);
/**
* @brief Same as rd_kafka_consume_start() but re-routes incoming messages to
* the provided queue \p rkqu (which must have been previously allocated
* with `rd_kafka_queue_new()`.
*
* The application must use one of the `rd_kafka_consume_*_queue()` functions
* to receive fetched messages.
*
* `rd_kafka_consume_start_queue()` must not be called multiple times for the
* same topic and partition without stopping consumption first with
* `rd_kafka_consume_stop()`.
* `rd_kafka_consume_start()` and `rd_kafka_consume_start_queue()` must not
* be combined for the same topic and partition.
*/
RD_EXPORT
int rd_kafka_consume_start_queue(rd_kafka_topic_t *rkt, int32_t partition,
int64_t offset, rd_kafka_queue_t *rkqu);
/**
* @brief Stop consuming messages for topic \p rkt and \p partition, purging
* all messages currently in the local queue.
*
* NOTE: To enforce synchronisation this call will block until the internal
* fetcher has terminated and offsets are committed to configured
* storage method.
*
* The application needs to be stop all consumers before calling
* `rd_kafka_destroy()` on the main object handle.
*
* @returns 0 on success or -1 on error (see `errno`).
*/
RD_EXPORT
int rd_kafka_consume_stop(rd_kafka_topic_t *rkt, int32_t partition);
/**
* @brief Seek consumer for topic+partition to \p offset which is either an
* absolute or logical offset.
*
* If \p timeout_ms is not 0 the call will wait this long for the
* seek to be performed. If the timeout is reached the internal state
* will be unknown and this function returns `RD_KAFKA_RESP_ERR__TIMED_OUT`.
* If \p timeout_ms is 0 it will initiate the seek but return
* immediately without any error reporting (e.g., async).
*
* This call triggers a fetch queue barrier flush.
*
* @returns `RD_KAFKA_RESP_ERR__NO_ERROR` on success else an error code.
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_seek (rd_kafka_topic_t *rkt,
int32_t partition,
int64_t offset,
int timeout_ms);
/**
* @brief Consume a single message from topic \p rkt and \p partition
*
* \p timeout_ms is maximum amount of time to wait for a message to be received.
* Consumer must have been previously started with `rd_kafka_consume_start()`.
*
* @returns a message object on success or \c NULL on error.
* The message object must be destroyed with `rd_kafka_message_destroy()`
* when the application is done with it.
*
* Errors (when returning NULL):
* - ETIMEDOUT - \p timeout_ms was reached with no new messages fetched.
* - ENOENT - \p rkt + \p partition is unknown.
* (no prior `rd_kafka_consume_start()` call)
*
* NOTE: The returned message's \c ..->err must be checked for errors.
* NOTE: \c ..->err \c == \c RD_KAFKA_RESP_ERR__PARTITION_EOF signals that the
* end of the partition has been reached, which should typically not be
* considered an error. The application should handle this case
* (e.g., ignore).
*
* @remark on_consume() interceptors may be called from this function prior to
* passing message to application.
*/
RD_EXPORT
rd_kafka_message_t *rd_kafka_consume(rd_kafka_topic_t *rkt, int32_t partition,
int timeout_ms);
/**
* @brief Consume up to \p rkmessages_size from topic \p rkt and \p partition
* putting a pointer to each message in the application provided
* array \p rkmessages (of size \p rkmessages_size entries).
*
* `rd_kafka_consume_batch()` provides higher throughput performance
* than `rd_kafka_consume()`.
*
* \p timeout_ms is the maximum amount of time to wait for all of
* \p rkmessages_size messages to be put into \p rkmessages.
* If no messages were available within the timeout period this function
* returns 0 and \p rkmessages remains untouched.
* This differs somewhat from `rd_kafka_consume()`.
*
* The message objects must be destroyed with `rd_kafka_message_destroy()`
* when the application is done with it.
*
* @returns the number of rkmessages added in \p rkmessages,
* or -1 on error (same error codes as for `rd_kafka_consume()`.
*
* @sa rd_kafka_consume()
*
* @remark on_consume() interceptors may be called from this function prior to
* passing message to application.
*/
RD_EXPORT
ssize_t rd_kafka_consume_batch(rd_kafka_topic_t *rkt, int32_t partition,
int timeout_ms,
rd_kafka_message_t **rkmessages,
size_t rkmessages_size);
/**
* @brief Consumes messages from topic \p rkt and \p partition, calling
* the provided callback for each consumed messsage.
*
* `rd_kafka_consume_callback()` provides higher throughput performance
* than both `rd_kafka_consume()` and `rd_kafka_consume_batch()`.
*
* \p timeout_ms is the maximum amount of time to wait for one or more messages
* to arrive.
*
* The provided \p consume_cb function is called for each message,
* the application \b MUST \b NOT call `rd_kafka_message_destroy()` on the
* provided \p rkmessage.
*
* The \p opaque argument is passed to the 'consume_cb' as \p opaque.
*
* @returns the number of messages processed or -1 on error.
*
* @sa rd_kafka_consume()
*
* @remark on_consume() interceptors may be called from this function prior to
* passing message to application.
*/
RD_EXPORT
int rd_kafka_consume_callback(rd_kafka_topic_t *rkt, int32_t partition,
int timeout_ms,
void (*consume_cb) (rd_kafka_message_t
*rkmessage,
void *opaque),
void *opaque);
/**
* @name Simple Consumer API (legacy): Queue consumers
* @{
*
* The following `..._queue()` functions are analogue to the functions above
* but reads messages from the provided queue \p rkqu instead.
* \p rkqu must have been previously created with `rd_kafka_queue_new()`
* and the topic consumer must have been started with
* `rd_kafka_consume_start_queue()` utilising the the same queue.
*/
/**
* @brief Consume from queue
*
* @sa rd_kafka_consume()
*/
RD_EXPORT
rd_kafka_message_t *rd_kafka_consume_queue(rd_kafka_queue_t *rkqu,
int timeout_ms);
/**
* @brief Consume batch of messages from queue
*
* @sa rd_kafka_consume_batch()
*/
RD_EXPORT
ssize_t rd_kafka_consume_batch_queue(rd_kafka_queue_t *rkqu,
int timeout_ms,
rd_kafka_message_t **rkmessages,
size_t rkmessages_size);
/**
* @brief Consume multiple messages from queue with callback
*
* @sa rd_kafka_consume_callback()
*/
RD_EXPORT
int rd_kafka_consume_callback_queue(rd_kafka_queue_t *rkqu,
int timeout_ms,
void (*consume_cb) (rd_kafka_message_t
*rkmessage,
void *opaque),
void *opaque);
/**@}*/
/**
* @name Simple Consumer API (legacy): Topic+partition offset store.
* @{
*
* If \c auto.commit.enable is true the offset is stored automatically prior to
* returning of the message(s) in each of the rd_kafka_consume*() functions
* above.
*/
/**
* @brief Store offset \p offset for topic \p rkt partition \p partition.
*
* The offset will be committed (written) to the offset store according
* to \c `auto.commit.interval.ms` or manual offset-less commit()
*
* @remark \c `enable.auto.offset.store` must be set to "false" when using this API.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success or an error code on error.
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_offset_store(rd_kafka_topic_t *rkt,
int32_t partition, int64_t offset);
/**
* @brief Store offsets for next auto-commit for one or more partitions.
*
* The offset will be committed (written) to the offset store according
* to \c `auto.commit.interval.ms` or manual offset-less commit().
*
* Per-partition success/error status propagated through each partition's
* \c .err field.
*
* @remark \c `enable.auto.offset.store` must be set to "false" when using this API.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success, or
* RD_KAFKA_RESP_ERR__UNKNOWN_PARTITION if none of the
* offsets could be stored, or
* RD_KAFKA_RESP_ERR__INVALID_ARG if \c enable.auto.offset.store is true.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_offsets_store(rd_kafka_t *rk,
rd_kafka_topic_partition_list_t *offsets);
/**@}*/
/**
* @name KafkaConsumer (C)
* @{
* @brief High-level KafkaConsumer C API
*
*
*
*/
/**
* @brief Subscribe to topic set using balanced consumer groups.
*
* Wildcard (regex) topics are supported:
* any topic name in the \p topics list that is prefixed with \c \"^\" will
* be regex-matched to the full list of topics in the cluster and matching
* topics will be added to the subscription list.
*
* The full topic list is retrieved every \c topic.metadata.refresh.interval.ms
* to pick up new or delete topics that match the subscription.
* If there is any change to the matched topics the consumer will
* immediately rejoin the group with the updated set of subscribed topics.
*
* Regex and full topic names can be mixed in \p topics.
*
* @remark Only the \c .topic field is used in the supplied \p topics list,
* all other fields are ignored.
*
* @remark subscribe() is an asynchronous method which returns immediately:
* background threads will (re)join the group, wait for group rebalance,
* issue any registered rebalance_cb, assign() the assigned partitions,
* and then start fetching messages. This cycle may take up to
* \c session.timeout.ms * 2 or more to complete.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success or
* RD_KAFKA_RESP_ERR__INVALID_ARG if list is empty, contains invalid
* topics or regexes.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_subscribe (rd_kafka_t *rk,
const rd_kafka_topic_partition_list_t *topics);
/**
* @brief Unsubscribe from the current subscription set.
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_unsubscribe (rd_kafka_t *rk);
/**
* @brief Returns the current topic subscription
*
* @returns An error code on failure, otherwise \p topic is updated
* to point to a newly allocated topic list (possibly empty).
*
* @remark The application is responsible for calling
* rd_kafka_topic_partition_list_destroy on the returned list.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_subscription (rd_kafka_t *rk,
rd_kafka_topic_partition_list_t **topics);
/**
* @brief Poll the consumer for messages or events.
*
* Will block for at most \p timeout_ms milliseconds.
*
* @remark An application should make sure to call consumer_poll() at regular
* intervals, even if no messages are expected, to serve any
* queued callbacks waiting to be called. This is especially
* important when a rebalance_cb has been registered as it needs
* to be called and handled properly to synchronize internal
* consumer state.
*
* @returns A message object which is a proper message if \p ->err is
* RD_KAFKA_RESP_ERR_NO_ERROR, or an event or error for any other
* value.
*
* @remark on_consume() interceptors may be called from this function prior to
* passing message to application.
*
* @remark When subscribing to topics the application must call poll at
* least every \c max.poll.interval.ms to remain a member of the
* consumer group.
*
* Noteworthy errors returned in \c ->err:
* - RD_KAFKA_RESP_ERR__MAX_POLL_EXCEEDED - application failed to call
* poll within `max.poll.interval.ms`.
*
* @sa rd_kafka_message_t
*/
RD_EXPORT
rd_kafka_message_t *rd_kafka_consumer_poll (rd_kafka_t *rk, int timeout_ms);
/**
* @brief Close down the KafkaConsumer.
*
* @remark This call will block until the consumer has revoked its assignment,
* calling the \c rebalance_cb if it is configured, committed offsets
* to broker, and left the consumer group.
* The maximum blocking time is roughly limited to session.timeout.ms.
*
* @returns An error code indicating if the consumer close was succesful
* or not.
*
* @remark The application still needs to call rd_kafka_destroy() after
* this call finishes to clean up the underlying handle resources.
*
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_consumer_close (rd_kafka_t *rk);
/**
* @brief Atomic assignment of partitions to consume.
*
* The new \p partitions will replace the existing assignment.
*
* When used from a rebalance callback the application shall pass the
* partition list passed to the callback (or a copy of it) (even if the list
* is empty) rather than NULL to maintain internal join state.
* A zero-length \p partitions will treat the partitions as a valid,
* albeit empty, assignment, and maintain internal state, while a \c NULL
* value for \p partitions will reset and clear the internal state.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_assign (rd_kafka_t *rk,
const rd_kafka_topic_partition_list_t *partitions);
/**
* @brief Returns the current partition assignment
*
* @returns An error code on failure, otherwise \p partitions is updated
* to point to a newly allocated partition list (possibly empty).
*
* @remark The application is responsible for calling
* rd_kafka_topic_partition_list_destroy on the returned list.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_assignment (rd_kafka_t *rk,
rd_kafka_topic_partition_list_t **partitions);
/**
* @brief Commit offsets on broker for the provided list of partitions.
*
* \p offsets should contain \c topic, \c partition, \c offset and possibly
* \c metadata.
* If \p offsets is NULL the current partition assignment will be used instead.
*
* If \p async is false this operation will block until the broker offset commit
* is done, returning the resulting success or error code.
*
* If a rd_kafka_conf_set_offset_commit_cb() offset commit callback has been
* configured the callback will be enqueued for a future call to
* rd_kafka_poll(), rd_kafka_consumer_poll() or similar.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_commit (rd_kafka_t *rk, const rd_kafka_topic_partition_list_t *offsets,
int async);
/**
* @brief Commit message's offset on broker for the message's partition.
*
* @sa rd_kafka_commit
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_commit_message (rd_kafka_t *rk, const rd_kafka_message_t *rkmessage,
int async);
/**
* @brief Commit offsets on broker for the provided list of partitions.
*
* See rd_kafka_commit for \p offsets semantics.
*
* The result of the offset commit will be posted on the provided \p rkqu queue.
*
* If the application uses one of the poll APIs (rd_kafka_poll(),
* rd_kafka_consumer_poll(), rd_kafka_queue_poll(), ..) to serve the queue
* the \p cb callback is required. \p opaque is passed to the callback.
*
* If using the event API the callback is ignored and the offset commit result
* will be returned as an RD_KAFKA_EVENT_COMMIT event. The \p opaque
* value will be available with rd_kafka_event_opaque()
*
* If \p rkqu is NULL a temporary queue will be created and the callback will
* be served by this call.
*
* @sa rd_kafka_commit()
* @sa rd_kafka_conf_set_offset_commit_cb()
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_commit_queue (rd_kafka_t *rk,
const rd_kafka_topic_partition_list_t *offsets,
rd_kafka_queue_t *rkqu,
void (*cb) (rd_kafka_t *rk,
rd_kafka_resp_err_t err,
rd_kafka_topic_partition_list_t *offsets,
void *opaque),
void *opaque);
/**
* @brief Retrieve committed offsets for topics+partitions.
*
* The \p offset field of each requested partition will either be set to
* stored offset or to RD_KAFKA_OFFSET_INVALID in case there was no stored
* offset for that partition.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success in which case the
* \p offset or \p err field of each \p partitions' element is filled
* in with the stored offset, or a partition specific error.
* Else returns an error code.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_committed (rd_kafka_t *rk,
rd_kafka_topic_partition_list_t *partitions,
int timeout_ms);
/**
* @brief Retrieve current positions (offsets) for topics+partitions.
*
* The \p offset field of each requested partition will be set to the offset
* of the last consumed message + 1, or RD_KAFKA_OFFSET_INVALID in case there was
* no previous message.
*
* @remark In this context the last consumed message is the offset consumed
* by the current librdkafka instance and, in case of rebalancing, not
* necessarily the last message fetched from the partition.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success in which case the
* \p offset or \p err field of each \p partitions' element is filled
* in with the stored offset, or a partition specific error.
* Else returns an error code.
*/
RD_EXPORT rd_kafka_resp_err_t
rd_kafka_position (rd_kafka_t *rk,
rd_kafka_topic_partition_list_t *partitions);
/**@}*/
/**
* @name Producer API
* @{
*
*
*/
/**
* @brief Producer message flags
*/
#define RD_KAFKA_MSG_F_FREE 0x1 /**< Delegate freeing of payload to rdkafka. */
#define RD_KAFKA_MSG_F_COPY 0x2 /**< rdkafka will make a copy of the payload. */
#define RD_KAFKA_MSG_F_BLOCK 0x4 /**< Block produce*() on message queue full.
* WARNING: If a delivery report callback
* is used the application MUST
* call rd_kafka_poll() (or equiv.)
* to make sure delivered messages
* are drained from the internal
* delivery report queue.
* Failure to do so will result
* in indefinately blocking on
* the produce() call when the
* message queue is full. */
#define RD_KAFKA_MSG_F_PARTITION 0x8 /**< produce_batch() will honor
* per-message partition. */
/**
* @brief Produce and send a single message to broker.
*
* \p rkt is the target topic which must have been previously created with
* `rd_kafka_topic_new()`.
*
* `rd_kafka_produce()` is an asynch non-blocking API.
* See `rd_kafka_conf_set_dr_msg_cb` on how to setup a callback to be called
* once the delivery status (success or failure) is known. The delivery report
* is trigged by the application calling `rd_kafka_poll()` (at regular
* intervals) or `rd_kafka_flush()` (at termination).
*
* Since producing is asynchronous, you should call `rd_kafka_flush()` before
* you destroy the producer. Otherwise, any outstanding messages will be
* silently discarded.
*
* When temporary errors occur, librdkafka automatically retries to produce the
* messages. Retries are triggered after retry.backoff.ms and when the
* leader broker for the given partition is available. Otherwise, librdkafka
* falls back to polling the topic metadata to monitor when a new leader is
* elected (see the topic.metadata.refresh.fast.interval.ms and
* topic.metadata.refresh.interval.ms configurations) and then performs a
* retry. A delivery error will occur if the message could not be produced
* within message.timeout.ms.
*
* See the "Message reliability" chapter in INTRODUCTION.md for more
* information.
*
* \p partition is the target partition, either:
* - RD_KAFKA_PARTITION_UA (unassigned) for
* automatic partitioning using the topic's partitioner function, or
* - a fixed partition (0..N)
*
* \p msgflags is zero or more of the following flags OR:ed together:
* RD_KAFKA_MSG_F_BLOCK - block \p produce*() call if
* \p queue.buffering.max.messages or
* \p queue.buffering.max.kbytes are exceeded.
* Messages are considered in-queue from the point they
* are accepted by produce() until their corresponding
* delivery report callback/event returns.
* It is thus a requirement to call
* rd_kafka_poll() (or equiv.) from a separate
* thread when F_BLOCK is used.
* See WARNING on \c RD_KAFKA_MSG_F_BLOCK above.
*
* RD_KAFKA_MSG_F_FREE - rdkafka will free(3) \p payload when it is done
* with it.
* RD_KAFKA_MSG_F_COPY - the \p payload data will be copied and the
* \p payload pointer will not be used by rdkafka
* after the call returns.
* RD_KAFKA_MSG_F_PARTITION - produce_batch() will honour per-message
* partition, either set manually or by the
* configured partitioner.
*
* .._F_FREE and .._F_COPY are mutually exclusive.
*
* If the function returns -1 and RD_KAFKA_MSG_F_FREE was specified, then
* the memory associated with the payload is still the caller's
* responsibility.
*
* \p payload is the message payload of size \p len bytes.
*
* \p key is an optional message key of size \p keylen bytes, if non-NULL it
* will be passed to the topic partitioner as well as be sent with the
* message to the broker and passed on to the consumer.
*
* \p msg_opaque is an optional application-provided per-message opaque
* pointer that will provided in the delivery report callback (`dr_cb`) for
* referencing this message.
*
* @remark on_send() and on_acknowledgement() interceptors may be called
* from this function. on_acknowledgement() will only be called if the
* message fails partitioning.
*
* @returns 0 on success or -1 on error in which case errno is set accordingly:
* - ENOBUFS - maximum number of outstanding messages has been reached:
* "queue.buffering.max.messages"
* (RD_KAFKA_RESP_ERR__QUEUE_FULL)
* - EMSGSIZE - message is larger than configured max size:
* "messages.max.bytes".
* (RD_KAFKA_RESP_ERR_MSG_SIZE_TOO_LARGE)
* - ESRCH - requested \p partition is unknown in the Kafka cluster.
* (RD_KAFKA_RESP_ERR__UNKNOWN_PARTITION)
* - ENOENT - topic is unknown in the Kafka cluster.
* (RD_KAFKA_RESP_ERR__UNKNOWN_TOPIC)
* - ECANCELED - fatal error has been raised on producer, see
* rd_kafka_fatal_error().
*
* @sa Use rd_kafka_errno2err() to convert `errno` to rdkafka error code.
*/
RD_EXPORT
int rd_kafka_produce(rd_kafka_topic_t *rkt, int32_t partition,
int msgflags,
void *payload, size_t len,
const void *key, size_t keylen,
void *msg_opaque);
/**
* @brief Produce and send a single message to broker.
*
* The message is defined by a va-arg list using \c rd_kafka_vtype_t
* tag tuples which must be terminated with a single \c RD_KAFKA_V_END.
*
* @returns \c RD_KAFKA_RESP_ERR_NO_ERROR on success, else an error code.
* \c RD_KAFKA_RESP_ERR__CONFLICT is returned if _V_HEADER and
* _V_HEADERS are mixed.
*
* @sa rd_kafka_produce, RD_KAFKA_V_END
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_producev (rd_kafka_t *rk, ...);
/**
* @brief Produce multiple messages.
*
* If partition is RD_KAFKA_PARTITION_UA the configured partitioner will
* be run for each message (slower), otherwise the messages will be enqueued
* to the specified partition directly (faster).
*
* The messages are provided in the array \p rkmessages of count \p message_cnt
* elements.
* The \p partition and \p msgflags are used for all provided messages.
*
* Honoured \p rkmessages[] fields are:
* - payload,len Message payload and length
* - key,key_len Optional message key
* - _private Message opaque pointer (msg_opaque)
* - err Will be set according to success or failure.
* Application only needs to check for errors if
* return value != \p message_cnt.
*
* @remark If \c RD_KAFKA_MSG_F_PARTITION is set in \p msgflags, the
* \c .partition field of the \p rkmessages is used instead of
* \p partition.
*
* @returns the number of messages succesfully enqueued for producing.
*
* @remark This interface does NOT support setting message headers on
* the provided \p rkmessages.
*/
RD_EXPORT
int rd_kafka_produce_batch(rd_kafka_topic_t *rkt, int32_t partition,
int msgflags,
rd_kafka_message_t *rkmessages, int message_cnt);
/**
* @brief Wait until all outstanding produce requests, et.al, are completed.
* This should typically be done prior to destroying a producer instance
* to make sure all queued and in-flight produce requests are completed
* before terminating.
*
* @remark This function will call rd_kafka_poll() and thus trigger callbacks.
*
* @returns RD_KAFKA_RESP_ERR__TIMED_OUT if \p timeout_ms was reached before all
* outstanding requests were completed, else RD_KAFKA_RESP_ERR_NO_ERROR
*
* @sa rd_kafka_outq_len()
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_flush (rd_kafka_t *rk, int timeout_ms);
/**
* @brief Purge messages currently handled by the producer instance.
*
* @param purge_flags tells which messages should be purged and how.
*
* The application will need to call rd_kafka_poll() or rd_kafka_flush()
* afterwards to serve the delivery report callbacks of the purged messages.
*
* Messages purged from internal queues fail with the delivery report
* error code set to RD_KAFKA_RESP_ERR__PURGE_QUEUE, while purged messages that
* are in-flight to or from the broker will fail with the error code set to
* RD_KAFKA_RESP_ERR__PURGE_INFLIGHT.
*
* @warning Purging messages that are in-flight to or from the broker
* will ignore any sub-sequent acknowledgement for these messages
* received from the broker, effectively making it impossible
* for the application to know if the messages were successfully
* produced or not. This may result in duplicate messages if the
* application retries these messages at a later time.
*
* @remark This call may block for a short time while background thread
* queues are purged.
*
* @returns RD_KAFKA_RESP_ERR_NO_ERROR on success,
* RD_KAFKA_RESP_ERR__INVALID_ARG if the \p purge flags are invalid
* or unknown,
* RD_KAFKA_RESP_ERR__NOT_IMPLEMENTED if called on a non-producer
* client instance.
*/
RD_EXPORT
rd_kafka_resp_err_t rd_kafka_purge (rd_kafka_t *rk, int purge_flags);
/**
* @brief Flags for rd_kafka_purge()
*/
/*!
* Purge messages in internal queues.
*/
#define RD_KAFKA_PURGE_F_QUEUE 0x1
/*!
* Purge messages in-flight to or from the broker.
* Purging these messages will void any future acknowledgements from the
* broker, making it impossible for the application to know if these
* messages were successfully delivered or not.
* Retrying these messages may lead to duplicates.
*/
#define RD_KAFKA_PURGE_F_INFLIGHT 0x2
/*!
* Don't wait for background thread queue purging to finish.
*/
#define RD_KAFKA_PURGE_F_NON_BLOCKING 0x4
/**@}*/
/**
* @name Metadata API
* @{
*
*
*/
/**
* @brief Broker information
*/
typedef struct rd_kafka_metadata_broker {
int32_t id; /**< Broker Id */
char *host; /**< Broker hostname */
int port; /**< Broker listening port */
} rd_kafka_metadata_broker_t;
/**
* @brief Partition information
*/
typedef struct rd_kafka_metadata_partition {
int32_t id; /**< Partition Id */
rd_kafka_resp_err_t err; /**< Partition error reported by broker */
int32_t leader; /**< Leader broker */
int replica_cnt; /**< Number of brokers in \p replicas */
int32_t *replicas; /**< Replica brokers */
int isr_cnt; /**< Number of ISR brokers in \p isrs */
int32_t *isrs; /**< In-Sync-Replica brokers */
} rd_kafka_metadata_partition_t;
/**
* @brief Topic information
*/
typedef struct rd_kafka_metadata_topic {
char *topic; /**< Topic name */
int partition_cnt; /**< Number of partitions in \p partitions*/
struct rd_kafka_metadata_partition *partitions; /**< Partitions */
rd_kafka_resp_err_t err; /**< Topic error reported by broker */
} rd_kafka_metadata_topic_t;
/**
* @brief Metadata container
*/
typedef struct rd_kafka_metadata {
int broker_cnt; /**< Number of brokers in \p brokers */
struct rd_kafka_metadata_broker *brokers; /**< Brokers */
int topic_cnt; /**< Number of topics in \p topics */
struct rd_kafka_metadata_topic *topics; /**< Topics */
int32_t orig_broker_id; /**< Broker originating this metadata */
char *orig_broker_name; /**< Name of originating broker */
} rd_kafka_metadata_t;
/**
* @brief Request Metadata from broker.
*
* Parameters:
* - \p all_topics if non-zero: request info about all topics in cluster,
* if zero: only request info about locally known topics.
* - \p only_rkt only request info about this topic
* - \p metadatap pointer to hold metadata result.
* The \p *metadatap pointer must be released
* with rd_kafka_metadata_destroy().
* - \p timeout_ms maximum response time before failing.
*
* Returns RD_KAFKA_RESP_ERR_NO_ERROR on success (in which case *metadatap)
* will be set, else RD_KAFKA_RESP_ERR__TIMED_OUT on timeout or
* other error code on error.
*/
RD_EXPORT
rd_kafka_resp_err_t
rd_kafka_metadata (rd_kafka_t *rk, int all_topics,
rd_kafka_topic_t *only_rkt,
const struct rd_kafka_metadata **metadatap,
int timeout_ms);
/**
* @brief Release metadata memory.
*/
RD_EXPORT
void rd_kafka_metadata_destroy(const struct rd_kafka_metadata *metadata);
/**@}*/
/**
* @name Client group information
* @{
*
*
*/
/**
* @brief Group member information
*
* For more information on \p member_metadata format, see
* https://cwiki.apache.org/confluence/display/KAFKA/A+Guide+To+The+Kafka+Protocol#AGuideToTheKafkaProtocol-GroupMembershipAPI
*
*/
struct rd_kafka_group_member_info {
char *member_id; /**< Member id (generated by broker) */
char *client_id; /**< Client's \p client.id */
char *client_host; /**< Client's hostname */
void *member_metadata; /**< Member metadata (binary),
* format depends on \p protocol_type. */
int member_metadata_size; /**< Member metadata size in bytes */
void *member_assignment; /**< Member assignment (binary),
* format depends on \p protocol_type. */
int member_assignment_size; /**< Member assignment size in bytes */
};
/**
* @brief Group information
*/
struct rd_kafka_group_info {
struct rd_kafka_metadata_broker broker; /**< Originating broker info */
char *group; /**< Group name */
rd_kafka_resp_err_t err; /**< Broker-originated error */
char *state; /**< Group state */
char *protocol_type; /**< Group protocol type */
char *protocol; /**< Group protocol */
struct rd_kafka_group_member_info *members; /**< Group members */
int member_cnt; /**< Group member count */
};
/**
* @brief List of groups
*
* @sa rd_kafka_group_list_destroy() to release list memory.
*/
struct rd_kafka_group_list {
struct rd_kafka_group_info *groups; /**< Groups */
int group_cnt; /**< Group count */
};
/**
* @brief List and describe client groups in cluster.
*
* \p group is an optional group name to describe, otherwise (\p NULL) all
* groups are returned.
*
* \p timeout_ms is the (approximate) maximum time to wait for response
* from brokers and must be a positive value.
*
* @returns \c RD_KAFKA_RESP_ERR__NO_ERROR on success and \p grplistp is
* updated to point to a newly allocated list of groups.
* \c RD_KAFKA_RESP_ERR__PARTIAL if not all brokers responded
* in time but at least one group is returned in \p grplistlp.
* \c RD_KAFKA_RESP_ERR__TIMED_OUT if no groups were returned in the
* given timeframe but not all brokers have yet responded, or
* if the list of brokers in the cluster could not be obtained within
* the given timeframe.
* \c RD_KAFKA_RESP_ERR__TRANSPORT if no brokers were found.
* Other error codes may also be returned from the request layer.
*
* The \p grplistp remains untouched if any error code is returned,
* with the exception of RD_KAFKA_RESP_ERR__PARTIAL which behaves
* as RD_KAFKA_RESP_ERR__NO_ERROR (success) but with an incomplete
* group list.
*
* @sa Use rd_kafka_group_list_destroy() to release list memory.