From fdf6461076e234905542d01ded7a55d017c449b5 Mon Sep 17 00:00:00 2001
From: Andrey Kovalev
-
+
00version.go
-
+
adminapi.go
-
+
adminoptions.go
-
+
build_glibc_linux.go
-
+
config.go
-
+
consumer.go
-
+
context.go
-
+
error.go
-
+
error_gen.go
-
+
event.go
-
+
generated_errors.go
-
+
handle.go
-
+
header.go
-
+
kafka.go
-
+
log.go
-
+
message.go
-
+
metadata.go
-
+
misc.go
-
+
+ mockcluster.go
+
+
offset.go
-
+
producer.go
-
+
testhelpers.go
-
+
time.go
@@ -1282,37 +1435,87 @@ possible complications with blocking Poll() calls.
LibrdkafkaLinkInfo explains how librdkafka was linked to the Go client
OffsetBeginning represents the earliest offset (logical)
OffsetEnd represents the latest offset (logical)
OffsetInvalid represents an invalid/unspecified offset
OffsetStored represents a stored offset
PartitionAny represents any partition (for partitioning),
or unspecified value (for all other cases)
LibraryVersion returns the underlying librdkafka library version as a
(version_int, version_str) tuple.
WriteErrorCodes writes Go error code constants to file from the
librdkafka error codes.
This function is not intended for public use.
+ ACLBinding specifies the operation and permission type for a specific principal
+over one or more resources of the same type. Used by `AdminClient.CreateACLs`,
+returned by `AdminClient.DescribeACLs` and `AdminClient.DeleteACLs`.
+
+ ACLBindingFilter specifies a filter used to return a list of ACL bindings matching some or all of its attributes.
+Used by `AdminClient.DescribeACLs` and `AdminClient.DeleteACLs`.
+
+ ACLBindingFilters is a slice of ACLBindingFilter that also implements
+the sort interface
+
+ ACLBindings is a slice of ACLBinding that also implements
+the sort interface
+
+ ACLOperation enumerates the different types of ACL operation.
+
+ ACLOperationFromString translates a ACL operation name to
+a ACLOperation value.
+
+ String returns the human-readable representation of an ACLOperation
+
+ ACLPermissionType enumerates the different types of ACL permission types.
+
+ ACLPermissionTypeFromString translates a ACL permission type name to
+a ACLPermissionType value.
+
+ String returns the human-readable representation of an ACLPermissionType
+
NewAdminClient creats a new AdminClient instance with a new underlying client instance
NewAdminClientFromConsumer derives a new AdminClient from an existing Consumer instance.
The AdminClient will use the same configuration and connections as the parent instance.
NewAdminClientFromProducer derives a new AdminClient from an existing Producer instance.
The AdminClient will use the same configuration and connections as the parent instance.
AlterConfigs alters/updates cluster resource configuration.
const (
// ResourceUnknown - Unknown
- ResourceUnknown = ResourceType(C.RD_KAFKA_RESOURCE_UNKNOWN)
+ ResourceUnknown = ResourceType(C.RD_KAFKA_RESOURCE_UNKNOWN)
// ResourceAny - match any resource type (DescribeConfigs)
- ResourceAny = ResourceType(C.RD_KAFKA_RESOURCE_ANY)
+ ResourceAny = ResourceType(C.RD_KAFKA_RESOURCE_ANY)
// ResourceTopic - Topic
- ResourceTopic = ResourceType(C.RD_KAFKA_RESOURCE_TOPIC)
+ ResourceTopic = ResourceType(C.RD_KAFKA_RESOURCE_TOPIC)
// ResourceGroup - Group
- ResourceGroup = ResourceType(C.RD_KAFKA_RESOURCE_GROUP)
+ ResourceGroup = ResourceType(C.RD_KAFKA_RESOURCE_GROUP)
// ResourceBroker - Broker
- ResourceBroker = ResourceType(C.RD_KAFKA_RESOURCE_BROKER)
+ ResourceBroker = ResourceType(C.RD_KAFKA_RESOURCE_BROKER)
)
const (
// ConfigSourceUnknown is the default value
- ConfigSourceUnknown = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_UNKNOWN_CONFIG)
+ ConfigSourceUnknown = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_UNKNOWN_CONFIG)
// ConfigSourceDynamicTopic is dynamic topic config that is configured for a specific topic
- ConfigSourceDynamicTopic = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_DYNAMIC_TOPIC_CONFIG)
+ ConfigSourceDynamicTopic = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_DYNAMIC_TOPIC_CONFIG)
// ConfigSourceDynamicBroker is dynamic broker config that is configured for a specific broker
- ConfigSourceDynamicBroker = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_DYNAMIC_BROKER_CONFIG)
+ ConfigSourceDynamicBroker = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_DYNAMIC_BROKER_CONFIG)
// ConfigSourceDynamicDefaultBroker is dynamic broker config that is configured as default for all brokers in the cluster
- ConfigSourceDynamicDefaultBroker = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_DYNAMIC_DEFAULT_BROKER_CONFIG)
+ ConfigSourceDynamicDefaultBroker = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_DYNAMIC_DEFAULT_BROKER_CONFIG)
// ConfigSourceStaticBroker is static broker config provided as broker properties at startup (e.g. from server.properties file)
- ConfigSourceStaticBroker = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_STATIC_BROKER_CONFIG)
+ ConfigSourceStaticBroker = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_STATIC_BROKER_CONFIG)
// ConfigSourceDefault is built-in default configuration for configs that have a default value
- ConfigSourceDefault = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_DEFAULT_CONFIG)
+ ConfigSourceDefault = ConfigSource(C.RD_KAFKA_CONFIG_SOURCE_DEFAULT_CONFIG)
+)
+ const (
+ // ResourcePatternTypeUnknown is a resource pattern type not known or not set.
+ ResourcePatternTypeUnknown = ResourcePatternType(C.RD_KAFKA_RESOURCE_PATTERN_UNKNOWN)
+ // ResourcePatternTypeAny matches any resource, used for lookups.
+ ResourcePatternTypeAny = ResourcePatternType(C.RD_KAFKA_RESOURCE_PATTERN_ANY)
+ // ResourcePatternTypeMatch will perform pattern matching
+ ResourcePatternTypeMatch = ResourcePatternType(C.RD_KAFKA_RESOURCE_PATTERN_MATCH)
+ // ResourcePatternTypeLiteral matches a literal resource name
+ ResourcePatternTypeLiteral = ResourcePatternType(C.RD_KAFKA_RESOURCE_PATTERN_LITERAL)
+ // ResourcePatternTypePrefixed matches a prefixed resource name
+ ResourcePatternTypePrefixed = ResourcePatternType(C.RD_KAFKA_RESOURCE_PATTERN_PREFIXED)
+)
+ const (
+ // ACLOperationUnknown represents an unknown or unset operation
+ ACLOperationUnknown = ACLOperation(C.RD_KAFKA_ACL_OPERATION_UNKNOWN)
+ // ACLOperationAny in a filter, matches any ACLOperation
+ ACLOperationAny = ACLOperation(C.RD_KAFKA_ACL_OPERATION_ANY)
+ // ACLOperationAll represents all the operations
+ ACLOperationAll = ACLOperation(C.RD_KAFKA_ACL_OPERATION_ALL)
+ // ACLOperationRead a read operation
+ ACLOperationRead = ACLOperation(C.RD_KAFKA_ACL_OPERATION_READ)
+ // ACLOperationWrite represents a write operation
+ ACLOperationWrite = ACLOperation(C.RD_KAFKA_ACL_OPERATION_WRITE)
+ // ACLOperationCreate represents a create operation
+ ACLOperationCreate = ACLOperation(C.RD_KAFKA_ACL_OPERATION_CREATE)
+ // ACLOperationDelete represents a delete operation
+ ACLOperationDelete = ACLOperation(C.RD_KAFKA_ACL_OPERATION_DELETE)
+ // ACLOperationAlter represents an alter operation
+ ACLOperationAlter = ACLOperation(C.RD_KAFKA_ACL_OPERATION_ALTER)
+ // ACLOperationDescribe represents a describe operation
+ ACLOperationDescribe = ACLOperation(C.RD_KAFKA_ACL_OPERATION_DESCRIBE)
+ // ACLOperationClusterAction represents a cluster action operation
+ ACLOperationClusterAction = ACLOperation(C.RD_KAFKA_ACL_OPERATION_CLUSTER_ACTION)
+ // ACLOperationDescribeConfigs represents a describe configs operation
+ ACLOperationDescribeConfigs = ACLOperation(C.RD_KAFKA_ACL_OPERATION_DESCRIBE_CONFIGS)
+ // ACLOperationAlterConfigs represents an alter configs operation
+ ACLOperationAlterConfigs = ACLOperation(C.RD_KAFKA_ACL_OPERATION_ALTER_CONFIGS)
+ // ACLOperationIdempotentWrite represents an idempotent write operation
+ ACLOperationIdempotentWrite = ACLOperation(C.RD_KAFKA_ACL_OPERATION_IDEMPOTENT_WRITE)
+)
+ const (
+ // ACLPermissionTypeUnknown represents an unknown ACLPermissionType
+ ACLPermissionTypeUnknown = ACLPermissionType(C.RD_KAFKA_ACL_PERMISSION_TYPE_UNKNOWN)
+ // ACLPermissionTypeAny in a filter, matches any ACLPermissionType
+ ACLPermissionTypeAny = ACLPermissionType(C.RD_KAFKA_ACL_PERMISSION_TYPE_ANY)
+ // ACLPermissionTypeDeny disallows access
+ ACLPermissionTypeDeny = ACLPermissionType(C.RD_KAFKA_ACL_PERMISSION_TYPE_DENY)
+ // ACLPermissionTypeAllow grants access
+ ACLPermissionTypeAllow = ACLPermissionType(C.RD_KAFKA_ACL_PERMISSION_TYPE_ALLOW)
)
const (
// TimestampNotAvailable indicates no timestamp was set, or not available due to lacking broker support
- TimestampNotAvailable = TimestampType(C.RD_KAFKA_TIMESTAMP_NOT_AVAILABLE)
+ TimestampNotAvailable = TimestampType(C.RD_KAFKA_TIMESTAMP_NOT_AVAILABLE)
// TimestampCreateTime indicates timestamp set by producer (source time)
- TimestampCreateTime = TimestampType(C.RD_KAFKA_TIMESTAMP_CREATE_TIME)
+ TimestampCreateTime = TimestampType(C.RD_KAFKA_TIMESTAMP_CREATE_TIME)
// TimestampLogAppendTime indicates timestamp set set by broker (store time)
- TimestampLogAppendTime = TimestampType(C.RD_KAFKA_TIMESTAMP_LOG_APPEND_TIME)
+ TimestampLogAppendTime = TimestampType(C.RD_KAFKA_TIMESTAMP_LOG_APPEND_TIME)
)
const (
// PurgeInFlight purges messages in-flight to or from the broker.
@@ -1320,75 +1523,253 @@ possible complications with blocking Poll() calls.
// broker, making it impossible for the application to know if these
// messages were successfully delivered or not.
// Retrying these messages may lead to duplicates.
- PurgeInFlight = int(C.RD_KAFKA_PURGE_F_INFLIGHT)
+ PurgeInFlight = int(C.RD_KAFKA_PURGE_F_INFLIGHT)
// PurgeQueue Purge messages in internal queues.
- PurgeQueue = int(C.RD_KAFKA_PURGE_F_QUEUE)
+ PurgeQueue = int(C.RD_KAFKA_PURGE_F_QUEUE)
// PurgeNonBlocking Don't wait for background thread queue purging to finish.
- PurgeNonBlocking = int(C.RD_KAFKA_PURGE_F_NON_BLOCKING)
+ PurgeNonBlocking = int(C.RD_KAFKA_PURGE_F_NON_BLOCKING)
)
const (
// AlterOperationSet sets/overwrites the configuration setting.
- AlterOperationSet = iota
+ AlterOperationSet = iota
)
const LibrdkafkaLinkInfo = "static glibc_linux from librdkafka-static-bundle-v1.8.2.tgz"
+ const LibrdkafkaLinkInfo = "static glibc_linux from librdkafka-static-bundle-v1.9.2.tgz"
const OffsetBeginning = Offset(C.RD_KAFKA_OFFSET_BEGINNING)
+ const OffsetBeginning = Offset(C.RD_KAFKA_OFFSET_BEGINNING)
const OffsetEnd = Offset(C.RD_KAFKA_OFFSET_END)
+ const OffsetEnd = Offset(C.RD_KAFKA_OFFSET_END)
const OffsetInvalid = Offset(C.RD_KAFKA_OFFSET_INVALID)
+ const OffsetInvalid = Offset(C.RD_KAFKA_OFFSET_INVALID)
const OffsetStored = Offset(C.RD_KAFKA_OFFSET_STORED)
+ const OffsetStored = Offset(C.RD_KAFKA_OFFSET_STORED)
const PartitionAny = int32(C.RD_KAFKA_PARTITION_UA)
+ const PartitionAny = int32(C.RD_KAFKA_PARTITION_UA)
func
-
+
LibraryVersion
¶
- func LibraryVersion() (int, string)
+ func LibraryVersion() (int, string)
func
-
+
WriteErrorCodes
¶
- func WriteErrorCodes(f *os.File)
+ func WriteErrorCodes(f *os.File)
+ type
+
+ ACLBinding
+
+
+ ¶
+
+
+ type ACLBinding struct {
+ Type ResourceType // The resource type.
+ // The resource name, which depends on the resource type.
+ // For ResourceBroker the resource name is the broker id.
+ Name string
+ ResourcePatternType ResourcePatternType // The resource pattern, relative to the name.
+ Principal string // The principal this ACLBinding refers to.
+ Host string // The host that the call is allowed to come from.
+ Operation ACLOperation // The operation/s specified by this binding.
+ PermissionType ACLPermissionType // The permission type for the specified operation.
+}
+
+
+ type
+
+ ACLBindingFilter
+
+
+ ¶
+
+
+ type ACLBindingFilter = ACLBinding
+
+ type
+
+ ACLBindingFilters
+
+
+ ¶
+
+
+ type ACLBindingFilters []ACLBindingFilter
+
+ type
+
+ ACLBindings
+
+
+ ¶
+
+
+ type ACLBindings []ACLBinding
+
+ func (ACLBindings)
+
+ Len
+
+
+ ¶
+
+
+ func (a ACLBindings) Len() int
+
+ func (ACLBindings)
+
+ Less
+
+
+ ¶
+
+
+ func (a ACLBindings) Less(i, j int) bool
+
+ func (ACLBindings)
+
+ Swap
+
+
+ ¶
+
+
+ func (a ACLBindings) Swap(i, j int)
+
+ type
+
+ ACLOperation
+
+
+ ¶
+
+
+ type ACLOperation int
+
+ func
+
+ ACLOperationFromString
+
+
+ ¶
+
+
+ func ACLOperationFromString(aclOperationString string) (ACLOperation, error)
+
+ func (ACLOperation)
+
+ String
+
+
+ ¶
+
+
+ func (o ACLOperation) String() string
+
+ type
+
+ ACLPermissionType
+
+
+ ¶
+
+
+ type ACLPermissionType int
+
+ func
+
+ ACLPermissionTypeFromString
+
+
+ ¶
+
+
+ func ACLPermissionTypeFromString(aclPermissionTypeString string) (ACLPermissionType, error)
+
+ func (ACLPermissionType)
+
+ String
+
+
+ ¶
+
+
+ func (o ACLPermissionType) String() string
+
type
-
+
AdminClient
@@ -1404,55 +1785,55 @@ This function is not intended for public use.
func
-
+
NewAdminClient
¶
- func NewAdminClient(conf *ConfigMap) (*AdminClient, error)
+ func NewAdminClient(conf *ConfigMap) (*AdminClient, error)
func
-
+
NewAdminClientFromConsumer
¶
- func NewAdminClientFromConsumer(c *Consumer) (a *AdminClient, err error)
+ func NewAdminClientFromConsumer(c *Consumer) (a *AdminClient, err error)
func
-
+
NewAdminClientFromProducer
¶
- func NewAdminClientFromProducer(p *Producer) (a *AdminClient, err error)
+ func NewAdminClientFromProducer(p *Producer) (a *AdminClient, err error)
func (*AdminClient)
-
+
AlterConfigs
¶
- func (a *AdminClient) AlterConfigs(ctx context.Context, resources []ConfigResource, options ...AlterConfigsAdminOption) (result []ConfigResourceResult, err error)
+ func (a *AdminClient) AlterConfigs(ctx context.Context, resources []ConfigResource, options ...AlterConfigsAdminOption) (result []ConfigResourceResult, err error)
func (a *AdminClient) ClusterID(ctx context.Context) (clusterID string, err error)+
func (a *AdminClient) ClusterID(ctx context.Context) (clusterID string, err error)
ClusterID returns the cluster ID as reported in broker metadata.
@@ -1512,14 +1893,14 @@ cancelling the context will block until the C function call returns.func (a *AdminClient) ControllerID(ctx context.Context) (controllerID int32, err error)+
func (a *AdminClient) ControllerID(ctx context.Context) (controllerID int32, err error)
ControllerID returns the broker ID of the current controller as reported in broker metadata. @@ -1532,29 +1913,53 @@ cancelling the context will block until the C function call returns.
Requires broker version >= 0.10.0.
+func (a *AdminClient) CreateACLs(ctx context.Context, aclBindings ACLBindings, options ...CreateACLsAdminOption) (result []CreateACLResult, err error)+
+ CreateACLs creates one or more ACL bindings. +
++ Parameters: +
+* `ctx` - context with the maximum amount of time to block, or nil for indefinite. +* `aclBindings` - A slice of ACL binding specifications to create. +* `options` - Create ACLs options ++
+ Returns a slice of CreateACLResult with a ErrNoError ErrorCode when the operation was successful +plus an error that is not nil for client level errors +
func (a *AdminClient) CreatePartitions(ctx context.Context, partitions []PartitionsSpecification, options ...CreatePartitionsAdminOption) (result []TopicResult, err error)+
func (a *AdminClient) CreatePartitions(ctx context.Context, partitions []PartitionsSpecification, options ...CreatePartitionsAdminOption) (result []TopicResult, err error)
CreatePartitions creates additional partitions for topics.
func (a *AdminClient) CreateTopics(ctx context.Context, topics []TopicSpecification, options ...CreateTopicsAdminOption) (result []TopicResult, err error)+
func (a *AdminClient) CreateTopics(ctx context.Context, topics []TopicSpecification, options ...CreateTopicsAdminOption) (result []TopicResult, err error)
CreateTopics creates topics in cluster.
@@ -1568,16 +1973,46 @@ make sure to check the result for topic-specific errors.Note: TopicSpecification is analogous to NewTopic in the Java Topic Admin API.
+func (a *AdminClient) DeleteACLs(ctx context.Context, aclBindingFilters ACLBindingFilters, options ...DeleteACLsAdminOption) (result []DeleteACLsResult, err error)+
+ DeleteACLs deletes ACL bindings matching one or more ACL binding filters. +
++ Parameters: +
+* `ctx` - context with the maximum amount of time to block, or nil for indefinite. +* `aclBindingFilters` - a slice of ACL binding filters to match ACLs to delete. + string attributes match exact values or any string if set to empty string. + Enum attributes match exact values or any value if ending with `Any`. + If `ResourcePatternType` is set to `ResourcePatternTypeMatch` deletes ACL bindings with: + - `ResourcePatternTypeLiteral` pattern type with resource name equal to the given resource name + - `ResourcePatternTypeLiteral` pattern type with wildcard resource name that matches the given resource name + - `ResourcePatternTypePrefixed` pattern type with resource name that is a prefix of the given resource name +* `options` - Delete ACLs options ++
+ Returns a slice of ACLBinding for each filter when the operation was successful +plus an error that is not `nil` for client level errors +
func (a *AdminClient) DeleteTopics(ctx context.Context, topics []string, options ...DeleteTopicsAdminOption) (result []TopicResult, err error)+
func (a *AdminClient) DeleteTopics(ctx context.Context, topics []string, options ...DeleteTopicsAdminOption) (result []TopicResult, err error)
DeleteTopics deletes a batch of topics.
@@ -1591,16 +2026,46 @@ topic metadata and configuration may continue to return information about deleteRequires broker version >= 0.10.1.0
+func (a *AdminClient) DescribeACLs(ctx context.Context, aclBindingFilter ACLBindingFilter, options ...DescribeACLsAdminOption) (result *DescribeACLsResult, err error)+
+ DescribeACLs matches ACL bindings by filter. +
++ Parameters: +
+* `ctx` - context with the maximum amount of time to block, or nil for indefinite. +* `aclBindingFilter` - A filter with attributes that must match. + string attributes match exact values or any string if set to empty string. + Enum attributes match exact values or any value if ending with `Any`. + If `ResourcePatternType` is set to `ResourcePatternTypeMatch` returns ACL bindings with: + - `ResourcePatternTypeLiteral` pattern type with resource name equal to the given resource name + - `ResourcePatternTypeLiteral` pattern type with wildcard resource name that matches the given resource name + - `ResourcePatternTypePrefixed` pattern type with resource name that is a prefix of the given resource name +* `options` - Describe ACLs options ++
+ Returns a slice of ACLBindings when the operation was successful +plus an error that is not `nil` for client level errors +
func (a *AdminClient) DescribeConfigs(ctx context.Context, resources []ConfigResource, options ...DescribeConfigsAdminOption) (result []ConfigResourceResult, err error)+
func (a *AdminClient) DescribeConfigs(ctx context.Context, resources []ConfigResource, options ...DescribeConfigsAdminOption) (result []ConfigResourceResult, err error)
DescribeConfigs retrieves configuration for cluster resources.
@@ -1633,14 +2098,14 @@ in the resource.func (a *AdminClient) GetMetadata(topic *string, allTopics bool, timeoutMs int) (*Metadata, error)+
func (a *AdminClient) GetMetadata(topic *string, allTopics bool, timeoutMs int) (*Metadata, error)
GetMetadata queries broker for cluster and topic metadata. If topic is non-nil only information about that topic is returned, else if @@ -1650,14 +2115,14 @@ GetMetadata is equivalent to listTopics, describeTopics and describeCluster in t
func (a *AdminClient) SetOAuthBearerToken(oauthBearerToken OAuthBearerToken) error+
func (a *AdminClient) SetOAuthBearerToken(oauthBearerToken OAuthBearerToken) error
SetOAuthBearerToken sets the the data to be transmitted to a broker during SASL/OAUTHBEARER authentication. It will return nil @@ -1675,14 +2140,14 @@ authentication mechanism.
func (a *AdminClient) SetOAuthBearerTokenFailure(errstr string) error+
func (a *AdminClient) SetOAuthBearerTokenFailure(errstr string) error
SetOAuthBearerTokenFailure sets the error message describing why token retrieval/setting failed; it also schedules a new token refresh event for 10 @@ -1694,20 +2159,20 @@ authentication mechanism.
func (a *AdminClient) String() string+
func (a *AdminClient) String() string
String returns a human readable name for an AdminClient instance
func SetAdminOperationTimeout(t time.Duration) (ao AdminOptionOperationTimeout)+
func SetAdminOperationTimeout(t time.Duration) (ao AdminOptionOperationTimeout)
SetAdminOperationTimeout sets the broker's operation timeout, such as the timeout for CreateTopics to complete the creation of topics on the controller @@ -1782,7 +2247,7 @@ in cluster.
func SetAdminRequestTimeout(t time.Duration) (ao AdminOptionRequestTimeout)+
func SetAdminRequestTimeout(t time.Duration) (ao AdminOptionRequestTimeout)
SetAdminRequestTimeout sets the overall request timeout, including broker lookup, request transmission, operation time on broker, and response. @@ -1825,7 +2290,7 @@ lookup, request transmission, operation time on broker, and response.
func SetAdminValidateOnly(validateOnly bool) (ao AdminOptionValidateOnly)+
func SetAdminValidateOnly(validateOnly bool) (ao AdminOptionValidateOnly)
SetAdminValidateOnly tells the broker to only validate the request, without performing the requested operation (create topics, etc). @@ -1868,7 +2333,7 @@ without performing the requested operation (create topics, etc).
type AlterOperation int+
type AlterOperation int
func (o AlterOperation) String() string+
func (o AlterOperation) String() string
String returns the human-readable representation of an AlterOperation
func (e AssignedPartitions) String() string+
func (e AssignedPartitions) String() string
type BrokerMetadata struct { - ID int32 - Host string - Port int + ID int32 + Host string + Port int }
type ConfigEntry struct { // Name of configuration entry, e.g., topic configuration property name. - Name string + Name string // Value of configuration entry. - Value string + Value string // Operation to perform on the entry. Operation AlterOperation }
func StringMapToConfigEntries(stringMap map[string]string, operation AlterOperation) []ConfigEntry+
func StringMapToConfigEntries(stringMap map[string]string, operation AlterOperation) []ConfigEntry
StringMapToConfigEntries creates a new map of ConfigEntry objects from the provided string map. The AlterOperation is set on each created entry.
func (c ConfigEntry) String() string+
func (c ConfigEntry) String() string
String returns a human-readable representation of a ConfigEntry.
type ConfigEntryResult struct { // Name of configuration entry, e.g., topic configuration property name. - Name string + Name string // Value of configuration entry. - Value string + Value string // Source indicates the configuration source. Source ConfigSource // IsReadOnly indicates whether the configuration entry can be altered. - IsReadOnly bool + IsReadOnly bool // IsSensitive indicates whether the configuration entry contains sensitive information, in which case the value will be unset. - IsSensitive bool + IsSensitive bool // IsSynonym indicates whether the configuration entry is a synonym for another configuration property. - IsSynonym bool + IsSynonym bool // Synonyms contains a map of configuration entries that are synonyms to this configuration entry. - Synonyms map[string]ConfigEntryResult + Synonyms map[string]ConfigEntryResult }
func (c ConfigEntryResult) String() string+
func (c ConfigEntryResult) String() string
String returns a human-readable representation of a ConfigEntryResult.
type ConfigMap map[string]ConfigValue+
type ConfigMap map[string]ConfigValue
func (m ConfigMap) Get(key string, defval ConfigValue) (ConfigValue, error)+
func (m ConfigMap) Get(key string, defval ConfigValue) (ConfigValue, error)
Get finds the given key in the ConfigMap and returns its value. If the key is not found `defval` is returned. @@ -2090,28 +2555,28 @@ an ErrInvalidArg error is returned.
func (m ConfigMap) Set(kv string) error+
func (m ConfigMap) Set(kv string) error
Set implements flag.Set (command line argument parser) as a convenience for `-X key=value` config.
func (m ConfigMap) SetKey(key string, value ConfigValue) error+
func (m ConfigMap) SetKey(key string, value ConfigValue) error
SetKey sets configuration property key to value.
@@ -2121,7 +2586,7 @@ set on the "default.topic.config" sub-map, this use is deprecated.func (c ConfigResource) String() string+
func (c ConfigResource) String() string
String returns a human-readable representation of a ConfigResource
func (c ConfigResourceResult) String() string+
func (c ConfigResourceResult) String() string
String returns a human-readable representation of a ConfigResourceResult.
ConfigSource represents an Apache Kafka config source
-type ConfigSource int+
type ConfigSource int
func (t ConfigSource) String() string+
func (t ConfigSource) String() string
String returns the human-readable representation of a ConfigSource type
type ConfigValue interface{}
func NewConsumer(conf *ConfigMap) (*Consumer, error)+
func NewConsumer(conf *ConfigMap) (*Consumer, error)
NewConsumer creates a new high-level Consumer instance.
@@ -2288,14 +2753,14 @@ event or message may be outdated.func (c *Consumer) Assign(partitions []TopicPartition) (err error)+
func (c *Consumer) Assign(partitions []TopicPartition) (err error)
Assign an atomic set of partitions to consume.
@@ -2311,27 +2776,27 @@ use the committed offset as a start position, with a fallback tofunc (c *Consumer) Assignment() (partitions []TopicPartition, err error)+
func (c *Consumer) Assignment() (partitions []TopicPartition, err error)
Assignment returns the current partition assignments
func (c *Consumer) AssignmentLost() bool+
func (c *Consumer) AssignmentLost() bool
AssignmentLost returns true if current partition assignment has been lost. This method is only applicable for use with a subscribing consumer when @@ -2341,28 +2806,28 @@ group and therefore commiting offsets, for example, may fail.
func (c *Consumer) Close() (err error)+
func (c *Consumer) Close() (err error)
Close Consumer instance. The object is no longer usable after this call.
func (c *Consumer) Commit() ([]TopicPartition, error)+
func (c *Consumer) Commit() ([]TopicPartition, error)
Commit offsets for currently assigned partitions This is a blocking call. @@ -2370,14 +2835,14 @@ Returns the committed offsets on success.
func (c *Consumer) CommitMessage(m *Message) ([]TopicPartition, error)+
func (c *Consumer) CommitMessage(m *Message) ([]TopicPartition, error)
CommitMessage commits offset based on the provided message. This is a blocking call. @@ -2385,14 +2850,14 @@ Returns the committed offsets on success.
func (c *Consumer) CommitOffsets(offsets []TopicPartition) ([]TopicPartition, error)+
func (c *Consumer) CommitOffsets(offsets []TopicPartition) ([]TopicPartition, error)
CommitOffsets commits the provided list of offsets This is a blocking call. @@ -2400,20 +2865,20 @@ Returns the committed offsets on success.
func (c *Consumer) Committed(partitions []TopicPartition, timeoutMs int) (offsets []TopicPartition, err error)+
func (c *Consumer) Committed(partitions []TopicPartition, timeoutMs int) (offsets []TopicPartition, err error)
Committed retrieves committed offsets for the given set of partitions
func (c *Consumer) GetConsumerGroupMetadata() (*ConsumerGroupMetadata, error)+
func (c *Consumer) GetConsumerGroupMetadata() (*ConsumerGroupMetadata, error)
GetConsumerGroupMetadata returns the consumer's current group metadata. This object should be passed to the transactional producer's @@ -2441,14 +2906,14 @@ SendOffsetsToTransaction() API.
func (c *Consumer) GetMetadata(topic *string, allTopics bool, timeoutMs int) (*Metadata, error)+
func (c *Consumer) GetMetadata(topic *string, allTopics bool, timeoutMs int) (*Metadata, error)
GetMetadata queries broker for cluster and topic metadata. If topic is non-nil only information about that topic is returned, else if @@ -2458,14 +2923,14 @@ GetMetadata is equivalent to listTopics, describeTopics and describeCluster in t
func (c *Consumer) GetRebalanceProtocol() string+
func (c *Consumer) GetRebalanceProtocol() string
GetRebalanceProtocol returns the current consumer group rebalance protocol, which is either "EAGER" or "COOPERATIVE". @@ -2475,14 +2940,14 @@ Should typically only be called during rebalancing.
func (c *Consumer) GetWatermarkOffsets(topic string, partition int32) (low, high int64, err error)+
func (c *Consumer) GetWatermarkOffsets(topic string, partition int32) (low, high int64, err error)
GetWatermarkOffsets returns the cached low and high offsets for the given topic and partition. The high offset is populated on every fetch response or via calling QueryWatermarkOffsets. @@ -2491,14 +2956,14 @@ OffsetInvalid will be returned if there is no cached offset for either value.
func (c *Consumer) IncrementalAssign(partitions []TopicPartition) (err error)+
func (c *Consumer) IncrementalAssign(partitions []TopicPartition) (err error)
IncrementalAssign adds the specified partitions to the current set of partitions to consume. @@ -2515,14 +2980,14 @@ use the committed offset as a start position, with a fallback to
func (c *Consumer) IncrementalUnassign(partitions []TopicPartition) (err error)+
func (c *Consumer) IncrementalUnassign(partitions []TopicPartition) (err error)
IncrementalUnassign removes the specified partitions from the current set of partitions to consume. @@ -2535,7 +3000,7 @@ partitions to consume.
func (c *Consumer) OffsetsForTimes(times []TopicPartition, timeoutMs int) (offsets []TopicPartition, err error)+
func (c *Consumer) OffsetsForTimes(times []TopicPartition, timeoutMs int) (offsets []TopicPartition, err error)
OffsetsForTimes looks up offsets by timestamp for the given partitions.
@@ -2579,14 +3044,14 @@ Per-partition errors may be returned in the `.Error` field.func (c *Consumer) Pause(partitions []TopicPartition) (err error)+
func (c *Consumer) Pause(partitions []TopicPartition) (err error)
Pause consumption for the provided list of partitions
@@ -2597,14 +3062,14 @@ this call, set `go.events.channel.size` accordingly.func (c *Consumer) Poll(timeoutMs int) (event Event)+
func (c *Consumer) Poll(timeoutMs int) (event Event)
Poll the consumer for messages or events.
@@ -2621,14 +3086,14 @@ this call, set `go.events.channel.size` accordingly.func (c *Consumer) Position(partitions []TopicPartition) (offsets []TopicPartition, err error)+
func (c *Consumer) Position(partitions []TopicPartition) (offsets []TopicPartition, err error)
Position returns the current consume position for the given partitions. Typical use is to call Assignment() to get the partition list @@ -2639,27 +3104,27 @@ i.e., the offset of the last message seen by the application + 1.
func (c *Consumer) QueryWatermarkOffsets(topic string, partition int32, timeoutMs int) (low, high int64, err error)+
func (c *Consumer) QueryWatermarkOffsets(topic string, partition int32, timeoutMs int) (low, high int64, err error)
QueryWatermarkOffsets queries the broker for the low and high offsets for the given topic and partition.
func (c *Consumer) ReadMessage(timeout time.Duration) (*Message, error)+
func (c *Consumer) ReadMessage(timeout time.Duration) (*Message, error)
ReadMessage polls the consumer for a message.
@@ -2686,27 +3151,27 @@ msg.TopicPartition provides partition-specific information (such as topic, partifunc (c *Consumer) Resume(partitions []TopicPartition) (err error)+
func (c *Consumer) Resume(partitions []TopicPartition) (err error)
Resume consumption for the provided list of partitions
func (c *Consumer) Seek(partition TopicPartition, timeoutMs int) error+
func (c *Consumer) Seek(partition TopicPartition, timeoutMs int) error
Seek seeks the given topic partitions using the offset from the TopicPartition.
@@ -2728,14 +3193,14 @@ a starting offset for each partition.func (c *Consumer) SetOAuthBearerToken(oauthBearerToken OAuthBearerToken) error+
func (c *Consumer) SetOAuthBearerToken(oauthBearerToken OAuthBearerToken) error
SetOAuthBearerToken sets the the data to be transmitted to a broker during SASL/OAUTHBEARER authentication. It will return nil @@ -2753,14 +3218,14 @@ authentication mechanism.
func (c *Consumer) SetOAuthBearerTokenFailure(errstr string) error+
func (c *Consumer) SetOAuthBearerTokenFailure(errstr string) error
SetOAuthBearerTokenFailure sets the error message describing why token retrieval/setting failed; it also schedules a new token refresh event for 10 @@ -2769,17 +3234,31 @@ success, otherwise an error if: 1) SASL/OAUTHBEARER is not supported by the underlying librdkafka build; 2) SASL/OAUTHBEARER is supported but is not configured as the client's authentication mechanism. +
+func (c *Consumer) StoreMessage(m *Message) (storedOffsets []TopicPartition, err error)+
+ StoreMessage stores offset based on the provided message. +This is a convenience method that uses StoreOffsets to do the actual work.
func (c *Consumer) StoreOffsets(offsets []TopicPartition) (storedOffsets []TopicPartition, err error)+
func (c *Consumer) StoreOffsets(offsets []TopicPartition) (storedOffsets []TopicPartition, err error)
StoreOffsets stores the provided list of offsets that will be committed to the offset store according to `auto.commit.interval.ms` or manual @@ -2792,87 +3271,87 @@ specific errors via its `.Error` member.
func (c *Consumer) String() string+
func (c *Consumer) String() string
Strings returns a human readable name for a Consumer instance
func (c *Consumer) Subscribe(topic string, rebalanceCb RebalanceCb) error+
func (c *Consumer) Subscribe(topic string, rebalanceCb RebalanceCb) error
Subscribe to a single topic This replaces the current subscription
func (c *Consumer) SubscribeTopics(topics []string, rebalanceCb RebalanceCb) (err error)+
func (c *Consumer) SubscribeTopics(topics []string, rebalanceCb RebalanceCb) (err error)
SubscribeTopics subscribes to the provided list of topics. This replaces the current subscription.
func (c *Consumer) Subscription() (topics []string, err error)+
func (c *Consumer) Subscription() (topics []string, err error)
Subscription returns the current subscription as set by Subscribe()
func (c *Consumer) Unassign() (err error)+
func (c *Consumer) Unassign() (err error)
Unassign the current set of partitions to consume.
func (c *Consumer) Unsubscribe() (err error)+
func (c *Consumer) Unsubscribe() (err error)
Unsubscribe from the current subscription, if any.
func NewTestConsumerGroupMetadata(groupID string) (*ConsumerGroupMetadata, error)+
func NewTestConsumerGroupMetadata(groupID string) (*ConsumerGroupMetadata, error)
NewTestConsumerGroupMetadata creates a new consumer group metadata instance mainly for testing use. Use GetConsumerGroupMetadata() to retrieve the real metadata.
++ CreateACLResult provides create ACL error information. +
+type CreateACLResult struct {
+ // Error, if any, of result. Check with `Error.Code() != ErrNoError`.
+ Error Error
+}
+
+ + CreateACLsAdminOption - see setter. +
++ See SetAdminRequestTimeout +
+type CreateACLsAdminOption interface {
+ // contains filtered or unexported methods
+}
type CreateTopicsAdminOption interface {
// contains filtered or unexported methods
}
+ + DeleteACLsAdminOption - see setter. +
++ See SetAdminRequestTimeout +
+type DeleteACLsAdminOption interface {
+ // contains filtered or unexported methods
+}
+ + DeleteACLsResult provides delete ACLs result or error information. +
+type DeleteACLsResult = DescribeACLsResult
type DeleteTopicsAdminOption interface {
// contains filtered or unexported methods
}
+ + DescribeACLsAdminOption - see setter. +
++ See SetAdminRequestTimeout +
+type DescribeACLsAdminOption interface {
+ // contains filtered or unexported methods
+}
+ + DescribeACLsResult provides describe ACLs result or error information. +
+type DescribeACLsResult struct { + // Slice of ACL bindings matching the provided filter + ACLBindings ACLBindings + // Error, if any, of result. Check with `Error.Code() != ErrNoError`. + Error Error +} +
func NewError(code ErrorCode, str string, fatal bool) (err Error)+
func NewError(code ErrorCode, str string, fatal bool) (err Error)
NewError creates a new Error.
func (e Error) Error() string+
func (e Error) Error() string
Error returns a human readable representation of an Error Same as Error.String()
func (e Error) IsFatal() bool+
func (e Error) IsFatal() bool
IsFatal returns true if the error is a fatal error. A fatal error indicates the client instance is no longer operable and @@ -3047,14 +3629,14 @@ idempotent producer errors.
func (e Error) IsRetriable() bool+
func (e Error) IsRetriable() bool
IsRetriable returns true if the operation that caused this error may be retried. @@ -3062,27 +3644,27 @@ This flag is currently only set by the Transactional producer API.
func (e Error) String() string+
func (e Error) String() string
String returns a human readable representation of an Error
func (e Error) TxnRequiresAbort() bool+
func (e Error) TxnRequiresAbort() bool
TxnRequiresAbort returns true if the error is an abortable transaction error that requires the application to abort the current transaction with @@ -3092,7 +3674,7 @@ This flag is only set by the Transactional producer API.
ErrorCode is the integer representation of local and broker error codes
-type ErrorCode int+
type ErrorCode int
const ( // ErrBadMsg Local: Bad message format - ErrBadMsg ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__BAD_MSG) + ErrBadMsg ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__BAD_MSG) // ErrBadCompression Local: Invalid compressed data - ErrBadCompression ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__BAD_COMPRESSION) + ErrBadCompression ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__BAD_COMPRESSION) // ErrDestroy Local: Broker handle destroyed - ErrDestroy ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__DESTROY) + ErrDestroy ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__DESTROY) // ErrFail Local: Communication failure with broker - ErrFail ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__FAIL) + ErrFail ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__FAIL) // ErrTransport Local: Broker transport failure - ErrTransport ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__TRANSPORT) + ErrTransport ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__TRANSPORT) // ErrCritSysResource Local: Critical system resource failure - ErrCritSysResource ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__CRIT_SYS_RESOURCE) + ErrCritSysResource ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__CRIT_SYS_RESOURCE) // ErrResolve Local: Host resolution failure - ErrResolve ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__RESOLVE) + ErrResolve ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__RESOLVE) // ErrMsgTimedOut Local: Message timed out - ErrMsgTimedOut ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__MSG_TIMED_OUT) + ErrMsgTimedOut ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__MSG_TIMED_OUT) // ErrPartitionEOF Broker: No more messages - ErrPartitionEOF ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PARTITION_EOF) + ErrPartitionEOF ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PARTITION_EOF) // ErrUnknownPartition Local: Unknown partition - ErrUnknownPartition ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_PARTITION) + ErrUnknownPartition ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_PARTITION) // ErrFs Local: File or filesystem error - ErrFs ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__FS) + ErrFs ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__FS) // ErrUnknownTopic Local: Unknown topic - ErrUnknownTopic ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_TOPIC) + ErrUnknownTopic ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_TOPIC) // ErrAllBrokersDown Local: All broker connections are down - ErrAllBrokersDown ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__ALL_BROKERS_DOWN) + ErrAllBrokersDown ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__ALL_BROKERS_DOWN) // ErrInvalidArg Local: Invalid argument or configuration - ErrInvalidArg ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__INVALID_ARG) + ErrInvalidArg ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__INVALID_ARG) // ErrTimedOut Local: Timed out - ErrTimedOut ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__TIMED_OUT) + ErrTimedOut ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__TIMED_OUT) // ErrQueueFull Local: Queue full - ErrQueueFull ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__QUEUE_FULL) + ErrQueueFull ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__QUEUE_FULL) // ErrIsrInsuff Local: ISR count insufficient - ErrIsrInsuff ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__ISR_INSUFF) + ErrIsrInsuff ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__ISR_INSUFF) // ErrNodeUpdate Local: Broker node update - ErrNodeUpdate ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NODE_UPDATE) + ErrNodeUpdate ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NODE_UPDATE) // ErrSsl Local: SSL error - ErrSsl ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__SSL) + ErrSsl ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__SSL) // ErrWaitCoord Local: Waiting for coordinator - ErrWaitCoord ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__WAIT_COORD) + ErrWaitCoord ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__WAIT_COORD) // ErrUnknownGroup Local: Unknown group - ErrUnknownGroup ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_GROUP) + ErrUnknownGroup ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_GROUP) // ErrInProgress Local: Operation in progress - ErrInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__IN_PROGRESS) + ErrInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__IN_PROGRESS) // ErrPrevInProgress Local: Previous operation in progress - ErrPrevInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PREV_IN_PROGRESS) + ErrPrevInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PREV_IN_PROGRESS) // ErrExistingSubscription Local: Existing subscription - ErrExistingSubscription ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__EXISTING_SUBSCRIPTION) + ErrExistingSubscription ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__EXISTING_SUBSCRIPTION) // ErrAssignPartitions Local: Assign partitions - ErrAssignPartitions ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__ASSIGN_PARTITIONS) + ErrAssignPartitions ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__ASSIGN_PARTITIONS) // ErrRevokePartitions Local: Revoke partitions - ErrRevokePartitions ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__REVOKE_PARTITIONS) + ErrRevokePartitions ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__REVOKE_PARTITIONS) // ErrConflict Local: Conflicting use - ErrConflict ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__CONFLICT) + ErrConflict ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__CONFLICT) // ErrState Local: Erroneous state - ErrState ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__STATE) + ErrState ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__STATE) // ErrUnknownProtocol Local: Unknown protocol - ErrUnknownProtocol ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_PROTOCOL) + ErrUnknownProtocol ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_PROTOCOL) // ErrNotImplemented Local: Not implemented - ErrNotImplemented ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NOT_IMPLEMENTED) + ErrNotImplemented ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NOT_IMPLEMENTED) // ErrAuthentication Local: Authentication failure - ErrAuthentication ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__AUTHENTICATION) + ErrAuthentication ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__AUTHENTICATION) // ErrNoOffset Local: No offset stored - ErrNoOffset ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NO_OFFSET) + ErrNoOffset ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NO_OFFSET) // ErrOutdated Local: Outdated - ErrOutdated ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__OUTDATED) + ErrOutdated ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__OUTDATED) // ErrTimedOutQueue Local: Timed out in queue - ErrTimedOutQueue ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__TIMED_OUT_QUEUE) + ErrTimedOutQueue ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__TIMED_OUT_QUEUE) // ErrUnsupportedFeature Local: Required feature not supported by broker - ErrUnsupportedFeature ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNSUPPORTED_FEATURE) + ErrUnsupportedFeature ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNSUPPORTED_FEATURE) // ErrWaitCache Local: Awaiting cache update - ErrWaitCache ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__WAIT_CACHE) + ErrWaitCache ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__WAIT_CACHE) // ErrIntr Local: Operation interrupted - ErrIntr ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__INTR) + ErrIntr ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__INTR) // ErrKeySerialization Local: Key serialization error - ErrKeySerialization ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__KEY_SERIALIZATION) + ErrKeySerialization ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__KEY_SERIALIZATION) // ErrValueSerialization Local: Value serialization error - ErrValueSerialization ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__VALUE_SERIALIZATION) + ErrValueSerialization ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__VALUE_SERIALIZATION) // ErrKeyDeserialization Local: Key deserialization error - ErrKeyDeserialization ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__KEY_DESERIALIZATION) + ErrKeyDeserialization ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__KEY_DESERIALIZATION) // ErrValueDeserialization Local: Value deserialization error - ErrValueDeserialization ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__VALUE_DESERIALIZATION) + ErrValueDeserialization ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__VALUE_DESERIALIZATION) // ErrPartial Local: Partial response - ErrPartial ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PARTIAL) + ErrPartial ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PARTIAL) // ErrReadOnly Local: Read-only object - ErrReadOnly ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__READ_ONLY) + ErrReadOnly ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__READ_ONLY) // ErrNoent Local: No such entry - ErrNoent ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NOENT) + ErrNoent ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NOENT) // ErrUnderflow Local: Read underflow - ErrUnderflow ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNDERFLOW) + ErrUnderflow ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNDERFLOW) // ErrInvalidType Local: Invalid type - ErrInvalidType ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__INVALID_TYPE) + ErrInvalidType ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__INVALID_TYPE) // ErrRetry Local: Retry operation - ErrRetry ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__RETRY) + ErrRetry ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__RETRY) // ErrPurgeQueue Local: Purged in queue - ErrPurgeQueue ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PURGE_QUEUE) + ErrPurgeQueue ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PURGE_QUEUE) // ErrPurgeInflight Local: Purged in flight - ErrPurgeInflight ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PURGE_INFLIGHT) + ErrPurgeInflight ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__PURGE_INFLIGHT) // ErrFatal Local: Fatal error - ErrFatal ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__FATAL) + ErrFatal ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__FATAL) // ErrInconsistent Local: Inconsistent state - ErrInconsistent ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__INCONSISTENT) + ErrInconsistent ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__INCONSISTENT) // ErrGaplessGuarantee Local: Gap-less ordering would not be guaranteed if proceeding - ErrGaplessGuarantee ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__GAPLESS_GUARANTEE) + ErrGaplessGuarantee ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__GAPLESS_GUARANTEE) // ErrMaxPollExceeded Local: Maximum application poll interval (max.poll.interval.ms) exceeded - ErrMaxPollExceeded ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__MAX_POLL_EXCEEDED) + ErrMaxPollExceeded ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__MAX_POLL_EXCEEDED) // ErrUnknownBroker Local: Unknown broker - ErrUnknownBroker ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_BROKER) + ErrUnknownBroker ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__UNKNOWN_BROKER) // ErrNotConfigured Local: Functionality not configured - ErrNotConfigured ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NOT_CONFIGURED) + ErrNotConfigured ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NOT_CONFIGURED) // ErrFenced Local: This instance has been fenced by a newer instance - ErrFenced ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__FENCED) + ErrFenced ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__FENCED) // ErrApplication Local: Application generated error - ErrApplication ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__APPLICATION) + ErrApplication ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__APPLICATION) // ErrAssignmentLost Local: Group partition assignment lost - ErrAssignmentLost ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__ASSIGNMENT_LOST) + ErrAssignmentLost ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__ASSIGNMENT_LOST) // ErrNoop Local: No operation performed - ErrNoop ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NOOP) + ErrNoop ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__NOOP) // ErrAutoOffsetReset Local: No offset to automatically reset to - ErrAutoOffsetReset ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__AUTO_OFFSET_RESET) + ErrAutoOffsetReset ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR__AUTO_OFFSET_RESET) // ErrUnknown Unknown broker error - ErrUnknown ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN) + ErrUnknown ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN) // ErrNoError Success - ErrNoError ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NO_ERROR) + ErrNoError ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NO_ERROR) // ErrOffsetOutOfRange Broker: Offset out of range - ErrOffsetOutOfRange ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OFFSET_OUT_OF_RANGE) + ErrOffsetOutOfRange ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OFFSET_OUT_OF_RANGE) // ErrInvalidMsg Broker: Invalid message - ErrInvalidMsg ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_MSG) + ErrInvalidMsg ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_MSG) // ErrUnknownTopicOrPart Broker: Unknown topic or partition - ErrUnknownTopicOrPart ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN_TOPIC_OR_PART) + ErrUnknownTopicOrPart ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN_TOPIC_OR_PART) // ErrInvalidMsgSize Broker: Invalid message size - ErrInvalidMsgSize ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_MSG_SIZE) + ErrInvalidMsgSize ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_MSG_SIZE) // ErrLeaderNotAvailable Broker: Leader not available - ErrLeaderNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_LEADER_NOT_AVAILABLE) + ErrLeaderNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_LEADER_NOT_AVAILABLE) // ErrNotLeaderForPartition Broker: Not leader for partition - ErrNotLeaderForPartition ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_LEADER_FOR_PARTITION) + ErrNotLeaderForPartition ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_LEADER_FOR_PARTITION) // ErrRequestTimedOut Broker: Request timed out - ErrRequestTimedOut ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_REQUEST_TIMED_OUT) + ErrRequestTimedOut ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_REQUEST_TIMED_OUT) // ErrBrokerNotAvailable Broker: Broker not available - ErrBrokerNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_BROKER_NOT_AVAILABLE) + ErrBrokerNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_BROKER_NOT_AVAILABLE) // ErrReplicaNotAvailable Broker: Replica not available - ErrReplicaNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_REPLICA_NOT_AVAILABLE) + ErrReplicaNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_REPLICA_NOT_AVAILABLE) // ErrMsgSizeTooLarge Broker: Message size too large - ErrMsgSizeTooLarge ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_MSG_SIZE_TOO_LARGE) + ErrMsgSizeTooLarge ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_MSG_SIZE_TOO_LARGE) // ErrStaleCtrlEpoch Broker: StaleControllerEpochCode - ErrStaleCtrlEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_STALE_CTRL_EPOCH) + ErrStaleCtrlEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_STALE_CTRL_EPOCH) // ErrOffsetMetadataTooLarge Broker: Offset metadata string too large - ErrOffsetMetadataTooLarge ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OFFSET_METADATA_TOO_LARGE) + ErrOffsetMetadataTooLarge ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OFFSET_METADATA_TOO_LARGE) // ErrNetworkException Broker: Broker disconnected before response received - ErrNetworkException ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NETWORK_EXCEPTION) + ErrNetworkException ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NETWORK_EXCEPTION) // ErrCoordinatorLoadInProgress Broker: Coordinator load in progress - ErrCoordinatorLoadInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_COORDINATOR_LOAD_IN_PROGRESS) + ErrCoordinatorLoadInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_COORDINATOR_LOAD_IN_PROGRESS) // ErrCoordinatorNotAvailable Broker: Coordinator not available - ErrCoordinatorNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_COORDINATOR_NOT_AVAILABLE) + ErrCoordinatorNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_COORDINATOR_NOT_AVAILABLE) // ErrNotCoordinator Broker: Not coordinator - ErrNotCoordinator ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_COORDINATOR) + ErrNotCoordinator ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_COORDINATOR) // ErrTopicException Broker: Invalid topic - ErrTopicException ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TOPIC_EXCEPTION) + ErrTopicException ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TOPIC_EXCEPTION) // ErrRecordListTooLarge Broker: Message batch larger than configured server segment size - ErrRecordListTooLarge ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_RECORD_LIST_TOO_LARGE) + ErrRecordListTooLarge ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_RECORD_LIST_TOO_LARGE) // ErrNotEnoughReplicas Broker: Not enough in-sync replicas - ErrNotEnoughReplicas ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_ENOUGH_REPLICAS) + ErrNotEnoughReplicas ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_ENOUGH_REPLICAS) // ErrNotEnoughReplicasAfterAppend Broker: Message(s) written to insufficient number of in-sync replicas - ErrNotEnoughReplicasAfterAppend ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_ENOUGH_REPLICAS_AFTER_APPEND) + ErrNotEnoughReplicasAfterAppend ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_ENOUGH_REPLICAS_AFTER_APPEND) // ErrInvalidRequiredAcks Broker: Invalid required acks value - ErrInvalidRequiredAcks ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_REQUIRED_ACKS) + ErrInvalidRequiredAcks ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_REQUIRED_ACKS) // ErrIllegalGeneration Broker: Specified group generation id is not valid - ErrIllegalGeneration ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_ILLEGAL_GENERATION) + ErrIllegalGeneration ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_ILLEGAL_GENERATION) // ErrInconsistentGroupProtocol Broker: Inconsistent group protocol - ErrInconsistentGroupProtocol ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INCONSISTENT_GROUP_PROTOCOL) + ErrInconsistentGroupProtocol ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INCONSISTENT_GROUP_PROTOCOL) // ErrInvalidGroupID Broker: Invalid group.id - ErrInvalidGroupID ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_GROUP_ID) + ErrInvalidGroupID ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_GROUP_ID) // ErrUnknownMemberID Broker: Unknown member - ErrUnknownMemberID ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN_MEMBER_ID) + ErrUnknownMemberID ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN_MEMBER_ID) // ErrInvalidSessionTimeout Broker: Invalid session timeout - ErrInvalidSessionTimeout ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_SESSION_TIMEOUT) + ErrInvalidSessionTimeout ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_SESSION_TIMEOUT) // ErrRebalanceInProgress Broker: Group rebalance in progress - ErrRebalanceInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_REBALANCE_IN_PROGRESS) + ErrRebalanceInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_REBALANCE_IN_PROGRESS) // ErrInvalidCommitOffsetSize Broker: Commit offset data size is not valid - ErrInvalidCommitOffsetSize ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_COMMIT_OFFSET_SIZE) + ErrInvalidCommitOffsetSize ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_COMMIT_OFFSET_SIZE) // ErrTopicAuthorizationFailed Broker: Topic authorization failed - ErrTopicAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TOPIC_AUTHORIZATION_FAILED) + ErrTopicAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TOPIC_AUTHORIZATION_FAILED) // ErrGroupAuthorizationFailed Broker: Group authorization failed - ErrGroupAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_GROUP_AUTHORIZATION_FAILED) + ErrGroupAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_GROUP_AUTHORIZATION_FAILED) // ErrClusterAuthorizationFailed Broker: Cluster authorization failed - ErrClusterAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_CLUSTER_AUTHORIZATION_FAILED) + ErrClusterAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_CLUSTER_AUTHORIZATION_FAILED) // ErrInvalidTimestamp Broker: Invalid timestamp - ErrInvalidTimestamp ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_TIMESTAMP) + ErrInvalidTimestamp ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_TIMESTAMP) // ErrUnsupportedSaslMechanism Broker: Unsupported SASL mechanism - ErrUnsupportedSaslMechanism ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSUPPORTED_SASL_MECHANISM) + ErrUnsupportedSaslMechanism ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSUPPORTED_SASL_MECHANISM) // ErrIllegalSaslState Broker: Request not valid in current SASL state - ErrIllegalSaslState ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_ILLEGAL_SASL_STATE) + ErrIllegalSaslState ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_ILLEGAL_SASL_STATE) // ErrUnsupportedVersion Broker: API version not supported - ErrUnsupportedVersion ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSUPPORTED_VERSION) + ErrUnsupportedVersion ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSUPPORTED_VERSION) // ErrTopicAlreadyExists Broker: Topic already exists - ErrTopicAlreadyExists ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TOPIC_ALREADY_EXISTS) + ErrTopicAlreadyExists ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TOPIC_ALREADY_EXISTS) // ErrInvalidPartitions Broker: Invalid number of partitions - ErrInvalidPartitions ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_PARTITIONS) + ErrInvalidPartitions ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_PARTITIONS) // ErrInvalidReplicationFactor Broker: Invalid replication factor - ErrInvalidReplicationFactor ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_REPLICATION_FACTOR) + ErrInvalidReplicationFactor ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_REPLICATION_FACTOR) // ErrInvalidReplicaAssignment Broker: Invalid replica assignment - ErrInvalidReplicaAssignment ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_REPLICA_ASSIGNMENT) + ErrInvalidReplicaAssignment ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_REPLICA_ASSIGNMENT) // ErrInvalidConfig Broker: Configuration is invalid - ErrInvalidConfig ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_CONFIG) + ErrInvalidConfig ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_CONFIG) // ErrNotController Broker: Not controller for cluster - ErrNotController ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_CONTROLLER) + ErrNotController ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NOT_CONTROLLER) // ErrInvalidRequest Broker: Invalid request - ErrInvalidRequest ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_REQUEST) + ErrInvalidRequest ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_REQUEST) // ErrUnsupportedForMessageFormat Broker: Message format on broker does not support request - ErrUnsupportedForMessageFormat ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSUPPORTED_FOR_MESSAGE_FORMAT) + ErrUnsupportedForMessageFormat ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSUPPORTED_FOR_MESSAGE_FORMAT) // ErrPolicyViolation Broker: Policy violation - ErrPolicyViolation ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_POLICY_VIOLATION) + ErrPolicyViolation ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_POLICY_VIOLATION) // ErrOutOfOrderSequenceNumber Broker: Broker received an out of order sequence number - ErrOutOfOrderSequenceNumber ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OUT_OF_ORDER_SEQUENCE_NUMBER) + ErrOutOfOrderSequenceNumber ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OUT_OF_ORDER_SEQUENCE_NUMBER) // ErrDuplicateSequenceNumber Broker: Broker received a duplicate sequence number - ErrDuplicateSequenceNumber ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DUPLICATE_SEQUENCE_NUMBER) + ErrDuplicateSequenceNumber ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DUPLICATE_SEQUENCE_NUMBER) // ErrInvalidProducerEpoch Broker: Producer attempted an operation with an old epoch - ErrInvalidProducerEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_PRODUCER_EPOCH) + ErrInvalidProducerEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_PRODUCER_EPOCH) // ErrInvalidTxnState Broker: Producer attempted a transactional operation in an invalid state - ErrInvalidTxnState ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_TXN_STATE) + ErrInvalidTxnState ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_TXN_STATE) // ErrInvalidProducerIDMapping Broker: Producer attempted to use a producer id which is not currently assigned to its transactional id - ErrInvalidProducerIDMapping ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_PRODUCER_ID_MAPPING) + ErrInvalidProducerIDMapping ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_PRODUCER_ID_MAPPING) // ErrInvalidTransactionTimeout Broker: Transaction timeout is larger than the maximum value allowed by the broker's max.transaction.timeout.ms - ErrInvalidTransactionTimeout ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_TRANSACTION_TIMEOUT) + ErrInvalidTransactionTimeout ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_TRANSACTION_TIMEOUT) // ErrConcurrentTransactions Broker: Producer attempted to update a transaction while another concurrent operation on the same transaction was ongoing - ErrConcurrentTransactions ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_CONCURRENT_TRANSACTIONS) + ErrConcurrentTransactions ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_CONCURRENT_TRANSACTIONS) // ErrTransactionCoordinatorFenced Broker: Indicates that the transaction coordinator sending a WriteTxnMarker is no longer the current coordinator for a given producer - ErrTransactionCoordinatorFenced ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TRANSACTION_COORDINATOR_FENCED) + ErrTransactionCoordinatorFenced ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TRANSACTION_COORDINATOR_FENCED) // ErrTransactionalIDAuthorizationFailed Broker: Transactional Id authorization failed - ErrTransactionalIDAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TRANSACTIONAL_ID_AUTHORIZATION_FAILED) + ErrTransactionalIDAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TRANSACTIONAL_ID_AUTHORIZATION_FAILED) // ErrSecurityDisabled Broker: Security features are disabled - ErrSecurityDisabled ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_SECURITY_DISABLED) + ErrSecurityDisabled ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_SECURITY_DISABLED) // ErrOperationNotAttempted Broker: Operation not attempted - ErrOperationNotAttempted ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OPERATION_NOT_ATTEMPTED) + ErrOperationNotAttempted ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OPERATION_NOT_ATTEMPTED) // ErrKafkaStorageError Broker: Disk error when trying to access log file on disk - ErrKafkaStorageError ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_KAFKA_STORAGE_ERROR) + ErrKafkaStorageError ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_KAFKA_STORAGE_ERROR) // ErrLogDirNotFound Broker: The user-specified log directory is not found in the broker config - ErrLogDirNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_LOG_DIR_NOT_FOUND) + ErrLogDirNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_LOG_DIR_NOT_FOUND) // ErrSaslAuthenticationFailed Broker: SASL Authentication failed - ErrSaslAuthenticationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_SASL_AUTHENTICATION_FAILED) + ErrSaslAuthenticationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_SASL_AUTHENTICATION_FAILED) // ErrUnknownProducerID Broker: Unknown Producer Id - ErrUnknownProducerID ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN_PRODUCER_ID) + ErrUnknownProducerID ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN_PRODUCER_ID) // ErrReassignmentInProgress Broker: Partition reassignment is in progress - ErrReassignmentInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_REASSIGNMENT_IN_PROGRESS) + ErrReassignmentInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_REASSIGNMENT_IN_PROGRESS) // ErrDelegationTokenAuthDisabled Broker: Delegation Token feature is not enabled - ErrDelegationTokenAuthDisabled ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_AUTH_DISABLED) + ErrDelegationTokenAuthDisabled ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_AUTH_DISABLED) // ErrDelegationTokenNotFound Broker: Delegation Token is not found on server - ErrDelegationTokenNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_NOT_FOUND) + ErrDelegationTokenNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_NOT_FOUND) // ErrDelegationTokenOwnerMismatch Broker: Specified Principal is not valid Owner/Renewer - ErrDelegationTokenOwnerMismatch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_OWNER_MISMATCH) + ErrDelegationTokenOwnerMismatch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_OWNER_MISMATCH) // ErrDelegationTokenRequestNotAllowed Broker: Delegation Token requests are not allowed on this connection - ErrDelegationTokenRequestNotAllowed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_REQUEST_NOT_ALLOWED) + ErrDelegationTokenRequestNotAllowed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_REQUEST_NOT_ALLOWED) // ErrDelegationTokenAuthorizationFailed Broker: Delegation Token authorization failed - ErrDelegationTokenAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_AUTHORIZATION_FAILED) + ErrDelegationTokenAuthorizationFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_AUTHORIZATION_FAILED) // ErrDelegationTokenExpired Broker: Delegation Token is expired - ErrDelegationTokenExpired ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_EXPIRED) + ErrDelegationTokenExpired ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DELEGATION_TOKEN_EXPIRED) // ErrInvalidPrincipalType Broker: Supplied principalType is not supported - ErrInvalidPrincipalType ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_PRINCIPAL_TYPE) + ErrInvalidPrincipalType ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_PRINCIPAL_TYPE) // ErrNonEmptyGroup Broker: The group is not empty - ErrNonEmptyGroup ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NON_EMPTY_GROUP) + ErrNonEmptyGroup ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NON_EMPTY_GROUP) // ErrGroupIDNotFound Broker: The group id does not exist - ErrGroupIDNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_GROUP_ID_NOT_FOUND) + ErrGroupIDNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_GROUP_ID_NOT_FOUND) // ErrFetchSessionIDNotFound Broker: The fetch session ID was not found - ErrFetchSessionIDNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_FETCH_SESSION_ID_NOT_FOUND) + ErrFetchSessionIDNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_FETCH_SESSION_ID_NOT_FOUND) // ErrInvalidFetchSessionEpoch Broker: The fetch session epoch is invalid - ErrInvalidFetchSessionEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_FETCH_SESSION_EPOCH) + ErrInvalidFetchSessionEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_FETCH_SESSION_EPOCH) // ErrListenerNotFound Broker: No matching listener - ErrListenerNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_LISTENER_NOT_FOUND) + ErrListenerNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_LISTENER_NOT_FOUND) // ErrTopicDeletionDisabled Broker: Topic deletion is disabled - ErrTopicDeletionDisabled ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TOPIC_DELETION_DISABLED) + ErrTopicDeletionDisabled ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_TOPIC_DELETION_DISABLED) // ErrFencedLeaderEpoch Broker: Leader epoch is older than broker epoch - ErrFencedLeaderEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_FENCED_LEADER_EPOCH) + ErrFencedLeaderEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_FENCED_LEADER_EPOCH) // ErrUnknownLeaderEpoch Broker: Leader epoch is newer than broker epoch - ErrUnknownLeaderEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN_LEADER_EPOCH) + ErrUnknownLeaderEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNKNOWN_LEADER_EPOCH) // ErrUnsupportedCompressionType Broker: Unsupported compression type - ErrUnsupportedCompressionType ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSUPPORTED_COMPRESSION_TYPE) + ErrUnsupportedCompressionType ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSUPPORTED_COMPRESSION_TYPE) // ErrStaleBrokerEpoch Broker: Broker epoch has changed - ErrStaleBrokerEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_STALE_BROKER_EPOCH) + ErrStaleBrokerEpoch ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_STALE_BROKER_EPOCH) // ErrOffsetNotAvailable Broker: Leader high watermark is not caught up - ErrOffsetNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OFFSET_NOT_AVAILABLE) + ErrOffsetNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_OFFSET_NOT_AVAILABLE) // ErrMemberIDRequired Broker: Group member needs a valid member ID - ErrMemberIDRequired ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_MEMBER_ID_REQUIRED) + ErrMemberIDRequired ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_MEMBER_ID_REQUIRED) // ErrPreferredLeaderNotAvailable Broker: Preferred leader was not available - ErrPreferredLeaderNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_PREFERRED_LEADER_NOT_AVAILABLE) + ErrPreferredLeaderNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_PREFERRED_LEADER_NOT_AVAILABLE) // ErrGroupMaxSizeReached Broker: Consumer group has reached maximum size - ErrGroupMaxSizeReached ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_GROUP_MAX_SIZE_REACHED) + ErrGroupMaxSizeReached ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_GROUP_MAX_SIZE_REACHED) // ErrFencedInstanceID Broker: Static consumer fenced by other consumer with same group.instance.id - ErrFencedInstanceID ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_FENCED_INSTANCE_ID) + ErrFencedInstanceID ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_FENCED_INSTANCE_ID) // ErrEligibleLeadersNotAvailable Broker: Eligible partition leaders are not available - ErrEligibleLeadersNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_ELIGIBLE_LEADERS_NOT_AVAILABLE) + ErrEligibleLeadersNotAvailable ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_ELIGIBLE_LEADERS_NOT_AVAILABLE) // ErrElectionNotNeeded Broker: Leader election not needed for topic partition - ErrElectionNotNeeded ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_ELECTION_NOT_NEEDED) + ErrElectionNotNeeded ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_ELECTION_NOT_NEEDED) // ErrNoReassignmentInProgress Broker: No partition reassignment is in progress - ErrNoReassignmentInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NO_REASSIGNMENT_IN_PROGRESS) + ErrNoReassignmentInProgress ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_NO_REASSIGNMENT_IN_PROGRESS) // ErrGroupSubscribedToTopic Broker: Deleting offsets of a topic while the consumer group is subscribed to it - ErrGroupSubscribedToTopic ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_GROUP_SUBSCRIBED_TO_TOPIC) + ErrGroupSubscribedToTopic ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_GROUP_SUBSCRIBED_TO_TOPIC) // ErrInvalidRecord Broker: Broker failed to validate record - ErrInvalidRecord ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_RECORD) + ErrInvalidRecord ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_RECORD) // ErrUnstableOffsetCommit Broker: There are unstable offsets that need to be cleared - ErrUnstableOffsetCommit ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSTABLE_OFFSET_COMMIT) + ErrUnstableOffsetCommit ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNSTABLE_OFFSET_COMMIT) // ErrThrottlingQuotaExceeded Broker: Throttling quota has been exceeded - ErrThrottlingQuotaExceeded ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_THROTTLING_QUOTA_EXCEEDED) + ErrThrottlingQuotaExceeded ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_THROTTLING_QUOTA_EXCEEDED) // ErrProducerFenced Broker: There is a newer producer with the same transactionalId which fences the current one - ErrProducerFenced ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_PRODUCER_FENCED) + ErrProducerFenced ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_PRODUCER_FENCED) // ErrResourceNotFound Broker: Request illegally referred to resource that does not exist - ErrResourceNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_RESOURCE_NOT_FOUND) + ErrResourceNotFound ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_RESOURCE_NOT_FOUND) // ErrDuplicateResource Broker: Request illegally referred to the same resource twice - ErrDuplicateResource ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DUPLICATE_RESOURCE) + ErrDuplicateResource ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_DUPLICATE_RESOURCE) // ErrUnacceptableCredential Broker: Requested credential would not meet criteria for acceptability - ErrUnacceptableCredential ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNACCEPTABLE_CREDENTIAL) + ErrUnacceptableCredential ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_UNACCEPTABLE_CREDENTIAL) // ErrInconsistentVoterSet Broker: Indicates that the either the sender or recipient of a voter-only request is not one of the expected voters - ErrInconsistentVoterSet ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INCONSISTENT_VOTER_SET) + ErrInconsistentVoterSet ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INCONSISTENT_VOTER_SET) // ErrInvalidUpdateVersion Broker: Invalid update version - ErrInvalidUpdateVersion ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_UPDATE_VERSION) + ErrInvalidUpdateVersion ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_INVALID_UPDATE_VERSION) // ErrFeatureUpdateFailed Broker: Unable to update finalized features due to server error - ErrFeatureUpdateFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_FEATURE_UPDATE_FAILED) + ErrFeatureUpdateFailed ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_FEATURE_UPDATE_FAILED) // ErrPrincipalDeserializationFailure Broker: Request principal deserialization failed during forwarding - ErrPrincipalDeserializationFailure ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_PRINCIPAL_DESERIALIZATION_FAILURE) + ErrPrincipalDeserializationFailure ErrorCode = ErrorCode(C.RD_KAFKA_RESP_ERR_PRINCIPAL_DESERIALIZATION_FAILURE) )
func (c ErrorCode) String() string+
func (c ErrorCode) String() string
String returns a human readable representation of an error code
type Event interface {
// String returns a human-readable representation of the event
- String() string
+ String() string
}
type Header struct { - Key string // Header name (utf-8 string) - Value []byte // Header value (nil, empty, or binary) + Key string // Header name (utf-8 string) + Value []byte // Header value (nil, empty, or binary) }
func (h Header) String() string+
func (h Header) String() string
String returns the Header Key and data in a human representable possibly truncated form suitable for displaying to the user.
type LogEvent struct { - Name string // Name of client instance - Tag string // Log tag that provides context to the log Message (e.g., "METADATA" or "GRPCOORD") - Message string // Log message - Level int // Log syslog level, lower is more critical. - Timestamp time.Time // Log timestamp + Name string // Name of client instance + Tag string // Log tag that provides context to the log Message (e.g., "METADATA" or "GRPCOORD") + Message string // Log message + Level int // Log syslog level, lower is more critical. + Timestamp time.Time // Log timestamp }
func (logEvent LogEvent) String() string+
func (logEvent LogEvent) String() string
type Message struct { TopicPartition TopicPartition - Value []byte - Key []byte - Timestamp time.Time + Value []byte + Key []byte + Timestamp time.Time TimestampType TimestampType Opaque interface{} Headers []Header @@ -3586,21 +4168,21 @@ suitable for displaying to the user.
func (m *Message) String() string+
func (m *Message) String() string
String returns a human readable representation of a Message. Key and payload are not represented.
type Metadata struct { Brokers []BrokerMetadata - Topics map[string]TopicMetadata + Topics map[string]TopicMetadata OriginatingBroker BrokerMetadata }+
+ MockCluster represents a Kafka mock cluster instance which can be used +for testing. +
+type MockCluster struct {
+ // contains filtered or unexported fields
+}
+
+ func NewMockCluster(brokerCount int) (*MockCluster, error)+
+ NewMockCluster provides a mock Kafka cluster with a configurable +number of brokers that support a reasonable subset of Kafka protocol +operations, error injection, etc. +
++ Mock clusters provide localhost listeners that can be used as the bootstrap +servers by multiple Kafka client instances. +
++ Currently supported functionality: +- Producer +- Idempotent Producer +- Transactional Producer +- Low-level consumer +- High-level balanced consumer groups with offset commits +- Topic Metadata and auto creation +
++ Warning THIS IS AN EXPERIMENTAL API, SUBJECT TO CHANGE OR REMOVAL. +
+func (mc *MockCluster) BootstrapServers() string+
+ BootstrapServers returns the bootstrap.servers property for this MockCluster +
+func (mc *MockCluster) Close()+
+ Close and destroy the MockCluster +
type OAuthBearerTokenRefresh struct {
// Config is the value of the sasl.oauthbearer.config property
- Config string
+ Config string
}
func (o OAuthBearerTokenRefresh) String() string+
func (o OAuthBearerTokenRefresh) String() string
Offset type (int64) with support for canonical names
-type Offset int64+
type Offset int64
func NewOffset(offset interface{}) (Offset, error)+
func NewOffset(offset interface{}) (Offset, error)
NewOffset creates a new Offset using the provided logical string, or an absolute int64 offset value. @@ -3706,7 +4362,7 @@ Logical offsets: "beginning", "earliest", "end", "latest", "unset", "invalid", "
func (o *Offset) Set(offset interface{}) error+
func (o *Offset) Set(offset interface{}) error
Set offset value, see NewOffset()
func (o Offset) String() string+
func (o Offset) String() string
type OffsetsCommitted struct { - Error error + Error error Offsets []TopicPartition }
func (o OffsetsCommitted) String() string+
func (o OffsetsCommitted) String() string
type PartitionEOF TopicPartition
func (p PartitionEOF) String() string+
func (p PartitionEOF) String() string
type PartitionMetadata struct { - ID int32 + ID int32 Error Error - Leader int32 - Replicas []int32 - Isrs []int32 + Leader int32 + Replicas []int32 + Isrs []int32 }
type PartitionsSpecification struct { // Topic to create more partitions for. - Topic string + Topic string // New partition count for topic, must be higher than current partition count. - IncreaseTo int + IncreaseTo int // (Optional) Explicit replica assignment. The outer array is // indexed by the new partition index (i.e., 0 for the first added // partition), while the inner per-partition array // contains the replica broker ids. The first broker in each // broker id list will be the preferred replica. - ReplicaAssignment [][]int32 + ReplicaAssignment [][]int32 }
func NewProducer(conf *ConfigMap) (*Producer, error)+
func NewProducer(conf *ConfigMap) (*Producer, error)
NewProducer creates a new high-level Producer instance.
@@ -3888,14 +4544,14 @@ go.logs.channel (chan kafka.LogEvent, nil) - Forward logs to application-providefunc (p *Producer) AbortTransaction(ctx context.Context) error+
func (p *Producer) AbortTransaction(ctx context.Context) error
AbortTransaction aborts the ongoing transaction.
@@ -3933,14 +4589,14 @@ has been raised by calling `err.(kafka.Error).IsFatal()`.func (p *Producer) BeginTransaction() error+
func (p *Producer) BeginTransaction() error
BeginTransaction starts a new transaction.
@@ -3970,7 +4626,7 @@ transaction, will fail.func (p *Producer) CommitTransaction(ctx context.Context) error+
func (p *Producer) CommitTransaction(ctx context.Context) error
CommitTransaction commits the current transaction.
@@ -4034,7 +4690,7 @@ respectively.func (p *Producer) Flush(timeoutMs int) int+
func (p *Producer) Flush(timeoutMs int) int
Flush and wait for outstanding messages and requests to complete delivery. Includes messages on ProduceChannel. @@ -4063,27 +4719,27 @@ Returns the number of outstanding events still un-flushed.
func (p *Producer) GetFatalError() error+
func (p *Producer) GetFatalError() error
GetFatalError returns an Error object if the client instance has raised a fatal error, else nil.
func (p *Producer) GetMetadata(topic *string, allTopics bool, timeoutMs int) (*Metadata, error)+
func (p *Producer) GetMetadata(topic *string, allTopics bool, timeoutMs int) (*Metadata, error)
GetMetadata queries broker for cluster and topic metadata. If topic is non-nil only information about that topic is returned, else if @@ -4093,14 +4749,14 @@ GetMetadata is equivalent to listTopics, describeTopics and describeCluster in t
func (p *Producer) InitTransactions(ctx context.Context) error+
func (p *Producer) InitTransactions(ctx context.Context) error
InitTransactions Initializes transactions for the producer instance.
@@ -4148,14 +4804,14 @@ error has been raised by calling `err.(kafka.Error).IsFatal()`.func (p *Producer) Len() int+
func (p *Producer) Len() int
Len returns the number of messages and requests waiting to be transmitted to the broker as well as delivery reports queued for the application. @@ -4163,7 +4819,7 @@ Includes messages on ProduceChannel.
func (p *Producer) OffsetsForTimes(times []TopicPartition, timeoutMs int) (offsets []TopicPartition, err error)+
func (p *Producer) OffsetsForTimes(times []TopicPartition, timeoutMs int) (offsets []TopicPartition, err error)
OffsetsForTimes looks up offsets by timestamp for the given partitions.
@@ -4207,14 +4863,14 @@ Per-partition errors may be returned in the `.Error` field.func (p *Producer) Produce(msg *Message, deliveryChan chan Event) error+
func (p *Producer) Produce(msg *Message, deliveryChan chan Event) error
Produce single message. This is an asynchronous call that enqueues the message on the internal @@ -4229,7 +4885,7 @@ Returns an error if message could not be enqueued.
func (p *Producer) Purge(flags int) error+
func (p *Producer) Purge(flags int) error
Purge messages currently handled by this producer instance.
@@ -4283,28 +4939,28 @@ queues are purged.func (p *Producer) QueryWatermarkOffsets(topic string, partition int32, timeoutMs int) (low, high int64, err error)+
func (p *Producer) QueryWatermarkOffsets(topic string, partition int32, timeoutMs int) (low, high int64, err error)
QueryWatermarkOffsets returns the broker's low and high offsets for the given topic and partition.
func (p *Producer) SendOffsetsToTransaction(ctx context.Context, offsets []TopicPartition, consumerMetadata *ConsumerGroupMetadata) error+
func (p *Producer) SendOffsetsToTransaction(ctx context.Context, offsets []TopicPartition, consumerMetadata *ConsumerGroupMetadata) error
SendOffsetsToTransaction sends a list of topic partition offsets to the consumer group coordinator for `consumerMetadata`, and marks the offsets @@ -4352,14 +5008,14 @@ respectively.
func (p *Producer) SetOAuthBearerToken(oauthBearerToken OAuthBearerToken) error+
func (p *Producer) SetOAuthBearerToken(oauthBearerToken OAuthBearerToken) error
SetOAuthBearerToken sets the the data to be transmitted to a broker during SASL/OAUTHBEARER authentication. It will return nil @@ -4377,14 +5033,14 @@ authentication mechanism.
func (p *Producer) SetOAuthBearerTokenFailure(errstr string) error+
func (p *Producer) SetOAuthBearerTokenFailure(errstr string) error
SetOAuthBearerTokenFailure sets the error message describing why token retrieval/setting failed; it also schedules a new token refresh event for 10 @@ -4396,34 +5052,34 @@ authentication mechanism.
func (p *Producer) String() string+
func (p *Producer) String() string
String returns a human readable name for a Producer instance
func (p *Producer) TestFatalError(code ErrorCode, str string) ErrorCode+
func (p *Producer) TestFatalError(code ErrorCode, str string) ErrorCode
TestFatalError triggers a fatal error in the underlying client. This is to be used strictly for testing purposes.
type RebalanceCb func(*Consumer, Event) error+
type RebalanceCb func(*Consumer, Event) error+
+ ResourcePatternType enumerates the different types of Kafka resource patterns. +
+type ResourcePatternType int+
func ResourcePatternTypeFromString(patternTypeString string) (ResourcePatternType, error)+
+ ResourcePatternTypeFromString translates a resource pattern type name to +a ResourcePatternType value. +
+func (t ResourcePatternType) String() string+
+ String returns the human-readable representation of a ResourcePatternType +
ResourceType represents an Apache Kafka resource type
-type ResourceType int+
type ResourceType int
func ResourceTypeFromString(typeString string) (ResourceType, error)+
func ResourceTypeFromString(typeString string) (ResourceType, error)
ResourceTypeFromString translates a resource type name/string to a ResourceType value.
func (t ResourceType) String() string+
func (t ResourceType) String() string
String returns the human-readable representation of a ResourceType
func (e RevokedPartitions) String() string+
func (e RevokedPartitions) String() string
func (e Stats) String() string+
func (e Stats) String() string
TimestampType is a the Message timestamp type or source
-type TimestampType int+
type TimestampType int
func (t TimestampType) String() string+
func (t TimestampType) String() string
type TopicMetadata struct { - Topic string + Topic string Partitions []PartitionMetadata Error Error }
type TopicPartition struct { - Topic *string - Partition int32 + Topic *string + Partition int32 Offset Offset - Metadata *string - Error error + Metadata *string + Error error }
func (p TopicPartition) String() string+
func (p TopicPartition) String() string
type TopicPartitions []TopicPartition
func (tps TopicPartitions) Len() int+
func (tps TopicPartitions) Len() int
func (tps TopicPartitions) Less(i, j int) bool+
func (tps TopicPartitions) Less(i, j int) bool
func (tps TopicPartitions) Swap(i, j int)+
func (tps TopicPartitions) Swap(i, j int)
type TopicResult struct { // Topic name - Topic string + Topic string // Error, if any, of result. Check with `Error.Code() != ErrNoError`. Error Error }
func (t TopicResult) String() string+
func (t TopicResult) String() string
String returns a human-readable representation of a TopicResult.
type TopicSpecification struct { // Topic name to create. - Topic string + Topic string // Number of partitions in topic. - NumPartitions int + NumPartitions int // Default replication factor for the topic's partitions, or zero // if an explicit ReplicaAssignment is set. - ReplicationFactor int + ReplicationFactor int // (Optional) Explicit replica assignment. The outer array is // indexed by the partition number, while the inner per-partition array // contains the replica broker ids. The first broker in each // broker id list will be the preferred replica. - ReplicaAssignment [][]int32 + ReplicaAssignment [][]int32 // Topic configuration. - Config map[string]string + Config map[string]string }diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin_amd64.go similarity index 52% rename from vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin.go rename to vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin_amd64.go index 5b6812f..96ba817 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin_amd64.go @@ -6,8 +6,8 @@ package kafka // #cgo CFLAGS: -DUSE_VENDORED_LIBRDKAFKA -DLIBRDKAFKA_STATICLIB -// #cgo LDFLAGS: ${SRCDIR}/librdkafka_vendor/librdkafka_darwin.a -lm -lsasl2 -ldl -lpthread +// #cgo LDFLAGS: ${SRCDIR}/librdkafka_vendor/librdkafka_darwin_amd64.a -lm -lsasl2 -ldl -lpthread -framework CoreFoundation -framework SystemConfiguration import "C" // LibrdkafkaLinkInfo explains how librdkafka was linked to the Go client -const LibrdkafkaLinkInfo = "static darwin from librdkafka-static-bundle-v1.8.2.tgz" +const LibrdkafkaLinkInfo = "static darwin_amd64 from librdkafka-static-bundle-v1.9.2.tgz" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin_arm64.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin_arm64.go new file mode 100644 index 0000000..d4d35c9 --- /dev/null +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_darwin_arm64.go @@ -0,0 +1,13 @@ +// +build !dynamic + + +// This file was auto-generated by librdkafka_vendor/bundle-import.sh, DO NOT EDIT. + +package kafka + +// #cgo CFLAGS: -DUSE_VENDORED_LIBRDKAFKA -DLIBRDKAFKA_STATICLIB +// #cgo LDFLAGS: ${SRCDIR}/librdkafka_vendor/librdkafka_darwin_arm64.a -lm -lsasl2 -ldl -lpthread -framework CoreFoundation -framework SystemConfiguration +import "C" + +// LibrdkafkaLinkInfo explains how librdkafka was linked to the Go client +const LibrdkafkaLinkInfo = "static darwin_arm64 from librdkafka-static-bundle-v1.9.2.tgz" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_glibc_linux.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_glibc_linux.go index a69638d..c30a134 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_glibc_linux.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_glibc_linux.go @@ -10,4 +10,4 @@ package kafka import "C" // LibrdkafkaLinkInfo explains how librdkafka was linked to the Go client -const LibrdkafkaLinkInfo = "static glibc_linux from librdkafka-static-bundle-v1.8.2.tgz" +const LibrdkafkaLinkInfo = "static glibc_linux from librdkafka-static-bundle-v1.9.2.tgz" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_musl_linux.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_musl_linux.go index ed2bee5..44569e0 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_musl_linux.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_musl_linux.go @@ -10,4 +10,4 @@ package kafka import "C" // LibrdkafkaLinkInfo explains how librdkafka was linked to the Go client -const LibrdkafkaLinkInfo = "static musl_linux from librdkafka-static-bundle-v1.8.2.tgz" +const LibrdkafkaLinkInfo = "static musl_linux from librdkafka-static-bundle-v1.9.2.tgz" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_windows.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_windows.go index 1c15c55..0727887 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_windows.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/build_windows.go @@ -10,4 +10,4 @@ package kafka import "C" // LibrdkafkaLinkInfo explains how librdkafka was linked to the Go client -const LibrdkafkaLinkInfo = "static windows from librdkafka-static-bundle-v1.8.2.tgz" +const LibrdkafkaLinkInfo = "static windows from librdkafka-static-bundle-v1.9.2.tgz" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/config.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/config.go index f0c8ecf..a6dfb8d 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/config.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/config.go @@ -1,5 +1,3 @@ -package kafka - /** * Copyright 2016 Confluent Inc. * @@ -16,6 +14,8 @@ package kafka * limitations under the License. */ +package kafka + import ( "fmt" "reflect" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/consumer.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/consumer.go index ef48330..2e1c9da 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/consumer.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/consumer.go @@ -1,5 +1,3 @@ -package kafka - /** * Copyright 2016-2020 Confluent Inc. * @@ -16,6 +14,8 @@ package kafka * limitations under the License. */ +package kafka + import ( "fmt" "math" @@ -46,7 +46,7 @@ type Consumer struct { readerTermChan chan bool rebalanceCb RebalanceCb appReassigned bool - appRebalanceEnable bool // Config setting + appRebalanceEnable bool // SerializerConfig setting } // Strings returns a human readable name for a Consumer instance @@ -424,30 +424,16 @@ func (c *Consumer) Close() (err error) { close(c.events) } - // librdkafka's rd_kafka_consumer_close() will block - // and trigger the rebalance_cb() if one is set, if not, which is the - // case with the Go client since it registers EVENTs rather than callbacks, - // librdkafka will shortcut the rebalance_cb and do a forced unassign. - // But we can't have that since the application might need the final RevokePartitions - // before shutting down. So we trigger an Unsubscribe() first, wait for that to - // propagate (in the Poll loop below), and then close the consumer. - c.Unsubscribe() + C.rd_kafka_consumer_close_queue(c.handle.rk, c.handle.rkq) - // Poll for rebalance events - for { - c.Poll(10 * 1000) - if int(C.rd_kafka_queue_length(c.handle.rkq)) == 0 { - break - } + for C.rd_kafka_consumer_closed(c.handle.rk) != 1 { + c.Poll(100) } // Destroy our queue C.rd_kafka_queue_destroy(c.handle.rkq) c.handle.rkq = nil - // Close the consumer - C.rd_kafka_consumer_close(c.handle.rk) - c.handle.cleanup() C.rd_kafka_destroy(c.handle.rk) diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/error.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/error.go index 1827f43..e14c1ca 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/error.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/error.go @@ -1,5 +1,3 @@ -package kafka - /** * Copyright 2016 Confluent Inc. * @@ -16,6 +14,8 @@ package kafka * limitations under the License. */ +package kafka + // Automatically generate error codes from librdkafka // See README for instructions //go:generate $GOPATH/bin/go_rdkafka_generr generated_errors.go @@ -67,10 +67,8 @@ func newCErrorFromString(code C.rd_kafka_resp_err_t, str string) (err Error) { return newErrorFromString(ErrorCode(code), str) } -// newErrorFromCError creates a new Error instance and destroys -// the passed cError. -func newErrorFromCErrorDestroy(cError *C.rd_kafka_error_t) Error { - defer C.rd_kafka_error_destroy(cError) +// newErrorFromCError creates a new Error instance +func newErrorFromCError(cError *C.rd_kafka_error_t) Error { return Error{ code: ErrorCode(C.rd_kafka_error_code(cError)), str: C.GoString(C.rd_kafka_error_string(cError)), @@ -80,6 +78,13 @@ func newErrorFromCErrorDestroy(cError *C.rd_kafka_error_t) Error { } } +// newErrorFromCErrorDestroy creates a new Error instance and destroys +// the passed cError. +func newErrorFromCErrorDestroy(cError *C.rd_kafka_error_t) Error { + defer C.rd_kafka_error_destroy(cError) + return newErrorFromCError(cError) +} + // Error returns a human readable representation of an Error // Same as Error.String() func (e Error) Error() string { diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/error_gen.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/error_gen.go index 4ceecb3..ccfaba2 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/error_gen.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/error_gen.go @@ -1,5 +1,3 @@ -package kafka - /** * Copyright 2020 Confluent Inc. * @@ -16,6 +14,8 @@ package kafka * limitations under the License. */ +package kafka + // Automatically generate error codes from librdkafka // See README for instructions //go:generate $GOPATH/bin/go_rdkafka_generr generated_errors.go diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/event.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/event.go index 6357ad8..aefa87b 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/event.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/event.go @@ -1,5 +1,3 @@ -package kafka - /** * Copyright 2016 Confluent Inc. * @@ -16,6 +14,8 @@ package kafka * limitations under the License. */ +package kafka + import ( "fmt" "os" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/generated_errors.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/generated_errors.go index fb828f9..d15e8f5 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/generated_errors.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/generated_errors.go @@ -1,6 +1,6 @@ package kafka -// Copyright 2016-2021 Confluent Inc. -// AUTOMATICALLY GENERATED ON 2021-12-08 12:44:39.243338672 +0100 CET m=+0.000248284 USING librdkafka 1.8.2 +// Copyright 2016-2022 Confluent Inc. +// AUTOMATICALLY GENERATED ON 2022-08-01 22:56:19.86222475 +0200 CEST m=+0.000294735 USING librdkafka 1.9.2 /* #include "select_rdkafka.h" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/handle.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/handle.go index 161a395..dc1c171 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/handle.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/handle.go @@ -1,5 +1,3 @@ -package kafka - /** * Copyright 2016 Confluent Inc. * @@ -16,6 +14,8 @@ package kafka * limitations under the License. */ +package kafka + import ( "fmt" "strings" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/header.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/header.go index f7a73a4..6baa381 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/header.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/header.go @@ -1,5 +1,3 @@ -package kafka - /** * Copyright 2018 Confluent Inc. * @@ -16,6 +14,8 @@ package kafka * limitations under the License. */ +package kafka + import ( "fmt" "strconv" diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/kafka.go b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/kafka.go index 20dc30a..254edbd 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/kafka.go +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/kafka.go @@ -20,7 +20,7 @@ // // High-level Consumer // -// * Decide if you want to read messages and events by calling `.Poll()` or +// * Decide if you want to read messages and events by calling `.Poll()` or // the deprecated option of using the `.Events()` channel. (If you want to use // `.Events()` channel then set `"go.events.channel.enable": true`). // diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/LICENSES.txt b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/LICENSES.txt index f2aa57d..1ab8a1d 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/LICENSES.txt +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/LICENSES.txt @@ -27,6 +27,32 @@ ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +LICENSE.cjson +-------------------------------------------------------------- +For cJSON.c and cJSON.h: + +Copyright (c) 2009-2017 Dave Gamble and cJSON contributors + +Permission is hereby granted, free of charge, to any person obtaining a copy +of this software and associated documentation files (the "Software"), to deal +in the Software without restriction, including without limitation the rights +to use, copy, modify, merge, publish, distribute, sublicense, and/or sell +copies of the Software, and to permit persons to whom the Software is +furnished to do so, subject to the following conditions: + +The above copyright notice and this permission notice shall be included in +all copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN +THE SOFTWARE. + + + LICENSE.crc32c -------------------------------------------------------------- # For src/crc32c.c copied (with modifications) from diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/bundle-import.sh b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/bundle-import.sh index 2f7aeda..9ad738f 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/bundle-import.sh +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/bundle-import.sh @@ -14,28 +14,30 @@ usage() { } +# Parse dynamic libraries from linker command line. +# Will print a list matching -lfoo and -framework X.. parse_dynlibs() { - # Parse dynamic libraries from pkg-config file, - # both the ones specified with Libs: but also through Requires: - local pc=$1 local libs= - local req= - local n= - for req in $(grep ^Requires: $pc | sed -e 's/^Requires://'); do - n=$(pkg-config --libs $req) - if [[ $n == -l* ]]; then - libs="${libs} $n" - fi - done - for n in $(grep ^Libs: $pc); do - if [[ $n == -l* ]]; then - libs="${libs} $n" + while [[ $# -gt 0 ]]; do + if [[ $1 == -l* ]]; then + libs="${libs} $1" + elif [[ $1 == -framework ]]; then + libs="${libs} $1 $2" + shift # remove one (extra) arg fi + shift # remove one arg done echo "$libs" } +# Parse dynamic library dependecies from pkg-config file and print +# them to stdout. +parse_pc_dynlibs() { + local pc=$1 + parse_dynlibs $(sed -n 's/^Libs: \(..*\)/\1/p' "$pc") +} + setup_build() { # Copies static library from the temp directory into final location, # extracts dynamic lib list from the pkg-config file, @@ -54,7 +56,7 @@ setup_build() { build_tag="// +build musl" fi - local dynlibs=$(parse_dynlibs $pc) + local dynlibs=$(parse_pc_dynlibs $pc) echo "Copying $apath to $dpath" cp "$apath" "$dpath" @@ -99,7 +101,7 @@ for f in rdkafka.h LICENSES.txt ; do done -for btype in glibc_linux musl_linux darwin windows ; do +for btype in glibc_linux musl_linux darwin_amd64 darwin_arm64 windows ; do lib=$bdir/librdkafka_${btype}.a pc=${lib/%.a/.pc} [[ -f $lib ]] || (echo "Expected file $lib missing" ; exit 1) diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/import.sh b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/import.sh index 7021169..7f920f3 100644 --- a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/import.sh +++ b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/import.sh @@ -51,12 +51,15 @@ fi curr_branch=$(git symbolic-ref HEAD 2>/dev/null | cut -d"/" -f 3-) uncommitted=$(git status --untracked-files=no --porcelain) -if [[ $devel != 1 ]] && ( [[ $curr_branch != master ]] || [[ ! -z $uncommitted ]] ); then +if [[ ! -z $uncommitted ]]; then + echo "Error: This script must be run on a clean branch with no uncommitted changes" + echo "Uncommitted files:" + echo "$uncommitted" + exit 1 +fi + +if [[ $devel != 1 ]] && [[ $curr_branch != master ]] ; then echo "Error: This script must be run on an up-to-date, clean, master branch" - if [[ ! -z $uncommitted ]]; then - echo "Uncommitted files:" - echo "$uncommitted" - fi exit 1 fi diff --git a/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/librdkafka_darwin.a b/vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/librdkafka_darwin_amd64.a similarity index 75% rename from vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/librdkafka_darwin.a rename to vendor/github.com/confluentinc/confluent-kafka-go/kafka/librdkafka_vendor/librdkafka_darwin_amd64.a index 6294180fee4e8b306a19d1f5d059572462c130b9..a49b5fae02198d53f8c95fb60afe88657240ff6e 100644 GIT binary patch delta 1710518 zcmXusbwE~07l82ty0&X~vBrvi1~ztg7XqS!BBEk~iWt`x8%(SfySrn}wR_dI6}!8z zzcb9^U%z|K%$=Eg@7#EUm&dbqCyqJyI&oC?q-DyNDp#&dnOO^m`(#U8x@1X@=qlNr zYUx~L(LGlQpEgT&
Slo zTjMZEqxFr2a5CkK@Dy9Pm-rms##@vRe5RB7trIG96Ba_{GZkJ@q2&v`x#S6xo*0Ge z@s1AItxuSI!i+j_6N|QIcKIIKP4UF;!7b1A=94^3KC@>}=^^oSa)rtrDt^KATG9V) zsN`eClXN(pL-3G2Hjq8`ojo>-J+^{+U#T~VdKt9b9I7?cd!dKMF~@48j%Y`Cs8nW` zKBU9{u|+AgK~rxb^?a$9QX4RPs3G;zvE}<2Kt}D@%$21KpeO@4%mDJe((~UmT%0Yf zjl1dSEDhzSp)E92g@!KCP&oz|%D{>+ur3VjEbVom{VCKN{6x=xGrOZGSc7}5OP21Ny1%ZDn+XUis?#$-iSX zo}ypDoUSj!*91u41$q 8tA0ou$`0eaaoxS-z<|mUWh*JM>fU z1zp7ZZ}YN0*pVXo)a%}+tJFQK&t(03NH;xJ!2zL?`h(rlX-V}PkT1ezTwYh{op`Xk zD&{Kw{|=MwWnE=Q&~WM3PhT4bO_D^lwO-T+nV^3IvOjc`bn?~uBS*`+cCPZc@fg_{ zsNbB}HCBrD(Z3P-F+tKM)k}1M(@3U{*1tv{Jy~W)nQt~sk<|KC3y+esWyC!FMuS(h z Z#4?pS^^J|#IW9m;DJDPY#g)f*vE0|+k5X)d2{aQ`h_({?o zhv9TwireuR-o{V(yJOoas5+*>>Lh%zH;%=XdI+-S87L`}=`JodQ2xXk=z~js)0@!W z;j$C^kv~HmMZ6YIC)4l$m PxDw|j*3)ZlxU9v!cm{V< z?=;@UcbHV?l9?l$b%1V%`Z==CF335&kMWr7OF~7NnbuXRy&ovca5v_l{4DVujK??z z_5oAqoG}}aJ3%L0N|LCFtuPR?GvXpx9a~@!{$x)UWk6N11$M$LDLIRbj!wQX3Mx=A zinu9pXB>(#xE>E<1175vULpSwKVkAzb^rxc$ISj}B)l*LN8%jZfamcgCP{4@%z|aG z2?k;WM&pXqod0`C+{SO1@ppC=mcl-nH9s;!hG8|$$MzT~t*|o=!ihKsPwILx3nOH? zxfEs{Jy5o5f$>vT{W{*A2uYF6`Uh6T#^{fOa2meKqO<#SgsdY@tnHcjF!3FX$K+0S z+ffemz^0g&2D%Xs$D<7RI)21-Y3zoq@X_0=UmMwjm+%RGM< d1N$=kPVA(FYH+1q0ISZMaT^MB*GQ$DUb9ydBTsbv?%Z zn Sg}rgA!@?Mj(PG?) zJvmm#h#%r8@^f%4Cd$O5#FBWJ^7_P`&>6(SB#gyvcpPuzPt2;Pl{qc{hdnu4hF}fy z&CwsDDPLrDWCw}!I9V$OPl%BDxZ12ZH9~f&9&;ijN^i@-Tf?OncR2@Ve$gkQ8~vp6 zcXe!k301Fb=`YL4@4ypyD?$DH|JNj3zS;)TV?Hd6wXqHMz@az=V{sE6!7KO#zZe}I zamsJD PX3L`3vLNyUa2(FZ^|&9;<6V5GI%bz8HVvzpu?VhJBk%W@eRv-4 z<7Z6u!`91<9#{igVGxGngdh6)-*ga5Vly7b%lHJpV(Oo^!#r3L>!L4);82XhMe3E- z17zJ#$1dz4aSSVu&{wUm+eu?J?qfT-sQIYi0rC`GL?@-j*|7xHMg291gz``vhq1U5 z&p9kS#BZ41#WvuMRnZ%J;b@$T8`b?W17ts*$2feYu2?-lzN3@T)m7Zi43IytI99<% z7^Fr$9U!4N9H-$j+=Zv{Hon2ciEKNWtorw#B+8?YRx|^etk3I@r-sWY&4(8Ymnod* zOL3<@cAFQa4Yd4m_HgmWVC;{J^zJxp)<9XUdi*s|wy1GS2FfXQ$989+9Mpwm8wbi% zvqI@`S%jN#AD+SM_}JJyT;AY!HP|OyPBE}JysGU-{yk9A=mtsmaLI+mjE*KMk%(u7 z$GU>|KLh10{=gLaipJz?m{&IU21+w*k0F>@HTCjhNsOR866f^M`+x5`6QzL;AlZ;` zX@gxb42R?Iy6Yp8M#$ATon&76hew$Fj*ZI~(LXA=HbiP+>~8&b?SG!K4Uggte6w4> z|6?lrPv4!|G1OBkVRH<^q gIRAf{)M~nEcV}{-~TZKctT>Re&ea0 zx4un=V`y*=uE+V5uf>CSn({2S^?QX6Ys>$zCbq@iIMQKZ5pKm3cnuHU({D%)uOt8A zOH6X#=5t^%Y>fWc4@YCY$NGL?p1R`ru@Ht6aUpKO3-}PfV#+7Bp~X4}%m)zL@i^YV zmne5_y$n|U{U;Lj(HBo$){2Q|N(k{d-2Fgj_k~$<9v|Si>iYiS?iR8DH{&fmB(dK- z a?!+^A6W?Rny|%#unCV~rZs^z=QUE=%8TP;_xK4G<3I|AB#gCZjf^Dz} z*25qihx2d?p2ydi=%Q^W2YO 0rVshiOcaMK0!~m^b)6AE8<7QoruFR)g{|; zZY+b%vC}2TR(MYXiS#;P&i@E1MAFbaT#GL_EhWyL=b6zHn_zDogDY?^#yKo}z~AoJ z2D4)f=j&=bfcNk-=6GPQ10}E}cEt%8i@7;%of0h6#f~@%=ipYnjBhagQ+xiqV@>Rg zgK#G9z<;fdydshInQf>5mca(t0sCN0J*~p043bW2wMBzufEu%WkhCU0OO0AHNXC(0 ztzOwUNDg3yzAhr$9bc)Hu;45C&gusP6=z6!Y>XW-94F#J+<~X@9>!zJbGF@uxAiWm zf3CX=&l??0Ow|Fzp7E7Y *+xmsD3i@Do9FMDU4_;Os)8Rc5 z?=bBI+i-ELfvqtZ2jN7VkE`*d`gn5@sXto3C;X(a_+w9eFhS3MQ{e*@lT5TLxM5Z7 zf>Uul9>a&2dXing9Ja(j9EuBZ_ar_4%>XWw7|zIIa2+1TEBGGMj ;Xkk}w!%=HiL3EAKEq^@ww;2QhxRL>(|`hBtVcobUwPtT z#396!i3bx;!3E@(6K^FxiTlW(wK{T@f|vN13O|Xz5@*&~ZzfU3DRvJPA+Ch&u@A1# zuMf4MgQctbI$*Gb;ZU5Epfgy0=z>fbEN9fX;e+KW-ov+;sDM6rOc^Z6RgdX|B@?>g zmk0L2B>h8cQLKeNc;tfp_3N8PXVnami#AaJ>tlNi!&$fo%jrj~p8o|(EzGi7zwwx> zvlPU#SOZ&P7mQFHv%*9YF}MP^ j0+sh(G1gYTT}AvJ}hT)MKo_mQx?w zVE3Xr2V95BK%9W{a7$6gRyauFJU+mWs$0IHlBKx*z`*TclBAf8Goq({h#X(JvHYgx z!B>Y#0rDmAck+L-enqTJc`Jtne+ DXRnSKU|(+qH;Fw zr0WfPKTMq5Br0M{?2FO36~kF^!mkD@>7F?{Tow>-!~^&*KE`jD&Ld#}j^woqQUWi~ z(F=6-v?H!gd3NH`SRXs%a9o5fS+5g@;n)P7VG={)Z&tX1FL4#+Nz2=T ^%4@V@S;*sH^(TSa|fSutEj8691iMPd^k#>@B=U9 zYwUt8D(mb2;LpQl0Ey0;F!2zKBtH+=VLV%MfcOf&z(ghOK(b*`tbuK?TS-0t%`Du& zE<1yda3~c&m)1*k-Qkj`f;|L(U Ku8>Tjn0Q`+6{koZpGB5^?)s*II> zb)W+=11^lUu{HI2U?c+`MjVX`aR)lbSh$66FtuJV%r5^E=hI*{; %cpnIMuVj{j=&gP zg}d=mf_nddOyY1}{Xnq7V7Y;xFr${6d_47j*XIG#Q9dj~d0nhbK529PksAL|a#>$+ zG}|#q%DmF&0`J-Kmu{H(371rf^xJVxU_+@-!6Pb0>Q|+bP3bK2F`j%X9gu104=ibJ zPVqye272KpIyj2HwBHlQVMi@@V($!*ay0m--p$+#HI^k**n_JW$mSWk1x=!)iSBZ@ z+MVS(9bBT`-_)xcs&nW4L}`b;aV*Zl6`^|mn~Hl#oW?l(h)Ke1gE_Di*2I<=j015J z&c|(d46lai`ELgBg2YdJ#+H4<)VgcUF>A!0%F3}Vgcb2Vdt!e-y-tjnB>&+XOxfS& z|HQ@)3*B%u#^Nr#gx@gR0NYS$Y=Xf!9OvS8JdcM4>s-1yNsoVotJK>&Np|36{D`Rs z+490z2Lo^@&chve5##AFW4Il#C$_ih?|+eq#`SmxpJ9rDwt+%e8{6P(It(Tb$MHBH z*Wtbdok@~XcZJyl=O`#h{G52vUENjs&gMoufM;>q7X89viy&Eo!#nEB`HDUgjSDcd z8eham3Sdd2qltA?*pKJ1eQkZmV($nEsiP06Nk&Q!;vrbGzJ7vHd!)2Lf6S)k!OKTT zQLK&8thW^RtB&d5UlMWn3X^HaraTL-kJaaU{js@@_4E}`;gM3Hs;mA={s^gzEiecj zoQbRO z2gd6Z^rM64#EGlfj&fk10y>bgO{KSPaZIC;vQ3R^IZ`|* zuZe#!@V^=Oc-r-%yl*wfHavmEV%(3{@f~KZZtJgQzz47h`Gs__3#*fFhv7Ju4R9Q` z;5b_lO@$R$oGqw^XW5c{lwZKd=u*QDFbfvJI@lgV(V4-*J+|aMrm1TWNikeTy$fvF z4&w9p1Q)VpQ`xXMHtZ+m8EV=A7PLCjkS*)nUjGE6iMQNh3*TT>R%nC0aS|@a{dfz% zV1`<@!}2)KN9Rn7(b70!!<>=Qku4mKORxzmJYkCy>6VyTp9=#iACB{J2YX- -b9bC=?;C>8*FTSG+;5PWnZK z);>~AV{@} @T6TEp z2MISu$sG#5Vun_BAlbD1*qKpM9$R8xJg1uV1L!Cm$KpKPh7YKJ(!`E#*-Z+*U~0WW znJsqFEjJBk!2DPm-)X%oKBJ`$aS;aW>TC1qupIf^#DUZws5+*DnIy8&@g%maCMz~Z zZ}gzvCf2J1LeAjPjwxB^f{!e9%VEe=~rEbU2xc^Yyeef0)pkiVt;?uIS+-5@^ke zi#fk*>shnE`Y36o_O3ljK5-!`LA@r}0sCX8rrMENA?}erVik{(c>Mjbje~C5Hx>ur z7@Uc#aW4+n_2d7DkvX^#yJ~x}6=GxvPQ|4T3!lvgkIOtI(N1ex{1Z!H6>Nxc7wvn; zA2IhO8`r{aI0bj$DRl0zko-TpVsUJZ5jgRoJ|$<2ku`V%pWru4ty^sB{ila6eCKF+ zh#xRXTbpmDzD_;HkpR0OeQ-3+!!39c@8V}n=9kcrWJeFIir&~8$KtXC_4Aj*B%Yva zJKNx4)uYoGIg2m-_1JD2BiD)F;1BX?+uQm%un3mJ#@Gpe5771;Z7`26ygo2S|LRH4 z+pA;5gSdh5%owRmJV -8QuR>t96T#JvV=|`?L14Q(1#*BMA z>iKV0xcI9AmhG;)G9*&I6Q}KD8!mvJ==9LngdXE$8dg>76&feW1NAMNm189X7Qn~4 zUR2Cj`Gj8r_55G)X{?l`LdwoIKB{r#ow0HQGm$ThWwAE4LVxUoBXKIm;sLzvu<#5& zVX7cIkoxS2ti*+|Jo;k2F1jcFh!ijEj6*OQSK&T%F0k+rzhb&z+h9qokN(&nC*lI! zf(P*$zQm+mZM!+G`u!ggHPIV;;uKtlf8zywg5R(^dn614*n>NGqPmFHLv*gZaU#Vp zVIfk+YsF+sBV`4q*2#1(=QwH3q#CL#ZqScQKhQu1t!Ez5j!^$07GS+DtT&zYMlm3F z?JwqOq?9r`x=@|T=Y`$T!C9=h3isi8yp4Tz0OqXuMx0s?rJ3A0aT@JC(vGjB9w%#6 z=XK(7vYdh *vJM*bv)dPaKDHHaT|1)g&J2 zJDBD-pC@U rDkn0UIyV9oQdaDkN-`OE9g9A;T^j6 zvK3NbRxE(!u!HJ7WrB3UemDXPXnR+Nj+2rYO@1~mw>q+g#1S0RTlc`_2{ISg;gddk z6B<2HK4PN2HXg?N2GN*up?#BThdxW1*ZGUcufrUapGYXz=l`_YP~5VK64Fq=3*M`h z x_(VB}k1-y*sixhbI2DtI+Vyf89c^F@1*Iuy`dNP}wP}6nhyyU` z7ZJC1^(8wN$3GJ32MPK)LV0X}oiH4y;W|94I%dVYBtGNsiEV?0u?Bi!Z;ZyJcoeVT zGjvPhBK}DlNd;_*9g~>f|ImUk5|eQm?!o`?K7PWX9J1*+`ltS!U$Tbs7d9=ae{8;@ zne@alc!9TQS|-)=-+X@GjnDB% 5nH6}}D+i}CkbWn 0@4ur>C h&a)`cb$T zx8r#{t*3L`(
ct2fzxm$9z^FF3-9rM5pDQzQ~89c zi`w|N!nVT#SPffaP6pDKcp9$8fABsg`CmeNj^wlpQU)F7LNu1A;o6u@^y429qXaAjsw3BgEQfUOeR)APRz4G+iZSe#wClz0yw#cTK(^QX6mqAGUAk+>Ak z;zOf3|5q{MFBBxrU>nMb#j!j#MSl##u{ayYvqu&YZ^YC1Pf0!h&F%3OiO)Ef6;o!k z1IdRau_1Oq2iLOR6yn9W4bNkJ4)xQFdj6XWP?AixqbyhqYhrs0#ee8%6!ARVj3@Ci zCdzEv&y6*(ZD#KO{Yi|*8+0&}coA;FlXxG$VY)1~qr6xVTVplmR0#1XhlLnihsW^_ zzQROVZ3pgH8(Uxx%)=FOH1Ry#fNg)BZO2IjFtR(u1Bu7u;zst`y$KKCWvoJn?^u6V zIejVT*<6xlvjcX+ve?k-NJkQ(I2 %?Q?H~1BkWj6!U z^FLF 1C)1_qvA nER7QVy8xoii2U=<9&p%{xB zb8-G3Cvg*hV7}b8!D`qTgE53zJdAh Ju(YeaPef)x9-$dMg)R*Va^o@p%G4chIJhyQMEQA%YDYnHvI1*>!LaYA& z4}IM)D<()1eVuR2i21QB*2jms;)ZkjQ4KEELuP&jcN-qW6CBbT_!6bAeL~ut@EMbS z|K}e*Y}z|szvokKiu{K!f5nt1onXuJU^!eqT0hZPG+EY-)YtnhCd*O0hA&Y@+49U- z$mr+_jY$OKFr0@Q@f^O!B-*j*FeiFoeGJ6G_=pYpivJs<@Bc;am@Kt)!I*ZK{n+iG zu3)yvk9 rz2e07=Oz-F >77JFecCSg|`jDeK9>wshboh(zyuf`XX_1&|; zDdHs2U11j7un#Ma#sjQy4}anu$}7@Awu$!aD2?@T9rXf;_Yn`Uu_MtGEW?9%3x8tv zNw$MhSRaFM2+qPycmm__1D;P9h@Stg*@EA7mzeY49c!Z>_Qz;kiTm&(Hf3@J<8X|@ zHMrmC=q$WI;u$86va>S}zT~t^s d63%#96o&|HiX;4}Vx4N%h_~ zlpD)oU2KCPI1*>#8a#pzG4Ti6Zg#AG$o|4|eH^PsopZX$q=bcT((bx`x;1a6^v5V% ziJ$K4S+Z%CBz<5#L_Qbs72=mzl6=w|w%x2)hrE~3(FTJ^6s1CCY>KBJ>t{a=V&pcy z$80aOf!i?>!TJ-h6#3fdhy8IPE>RuRK(Z&cgVi+9nT`kIHdZ)|_i!TR3vmY?!)utc zkG``S N9f_Xk;AH#@*W>tE`YG4fC>i;WtGK0}F0sUW@dnPJd>wuv zpE1U+?-8TtziH?Y75>8pR0zaFCtao5wdryhpWxmT`rxs1x}3oqnEtSS^qO_L{E42} z5CgEk!@}%&dh-d3mJQf(wtg6$Gg=1XLGlgP>uES+s$Acz&wlzBkds({F0R4@_}@-# zH*}i3#uU43oFARaEL^68C+Irg4k!nf$L83Q^{&xTF!3mS&H&$1K9&vnNj%+LM^;5S zvcxXPcAP^4TQKzzyDQHS7bbp8Ec@)9NP|Cqt+(55X$|VT9M!vKqv>*Ot$r}+OqKWq zbs-Z4g)s9vTcH#-!YY(E$1dbY;e5=Yw^{FvzVbgzv_QY-Gcj7SVHs>`bhMy5iE6qB z{9~eI2=NqLgN3!+d$3Mc;v-A-w0#sUH!&VFEVKEt*a64mdOV>zrhz9Uk}bC@=Em~a z3I|{;p2XLfa)n*LFkWJy4>3`!-4hq|(7YZyP2TR-`@dO8x5p;7AJwN-@9A<1J;*n} zyW~G&5cxqk12^Ikyo*0@7Lzgi-;V7#?>;+;QXaDdDEYU|XD2R<^~tx#2%L! ~9(mtTyO!usfkVK~X^ z$Wjvf@e01c-=5kAQ|hHPe))9CgC(%%3LVhzQ)L9s#~pY9pJVcswtj&G_53f*v9Cpg z6{*mIxFvB{{K(yJ1o1pvkH_&Ie#305Y{w lu@V;IbZJLii+C_`2jYP^17}gb1^>l&m~o@+ue`%TYwV9v zxB*Y$YfQDtHt;`egdw bTWL|6=E{63n z2uI-TU(VIZSua{P6aRx3@fjvwV%yD)717&h?*9WxY@xwu;^nvxFX0PJruTdEU{L_8 zqYw7NF&K-R@zxqW|E0lHd8rFVSA8ThX23#N6 9FEcWEL6WW7w;=yFlm^LkL#iI zs60&`W0Gw)&bLj^e^XF{gfI5RDYyau#aEbWyKSftHo)#U78l_jyov9(>-ld6kZOmu zFxJJ6I2`}NeRv0dV2+)(ol4ja2V)FwMIVPt;Wu7TWIE-V@IoVmd{N>mD6{ksosW`; zz52>$+cY^b+rCb?hgqK5yQMp(d}&`J=E7oF8J%)e{73^0h*xq &jibLJ)3 zCfW&Oy{^<7PaMF&`r@2~yx#x+)zi+qNwmDg 2Z2exeGnzwo zTI)OJMdlM$Orm?h#5u7PR;8m6?9yr2Vy1n;(jA9kJ?f`rvVK0J1IRx^+MU)5R+1SK zfib7`{5OehBramAvv!4##2K^!Q?C?yV?H`4j#aQ3&R|QIFyLJ5i4xf2q@Mp~^3)+w z{h2)!Ug$@@2adwIxRMUGU z6#EBPItZjDxN1c z!!9@kr{YT7hgb0hCOdE2D}7b>beWlsG`9=V4Ts@O+<^b!Z9K*Ruj4ax)n_~N`aT<$ z!9u6?BU}Anll3qtK_^6pl1Mg6-!S~Au%tnEERT)Q7rWyqjK&4H4ENzVyoYa$j&|s( zKPG27&VlY&7K`7pzX@3j126*T;|4s6SMeimeW1UbroSP30w1Z4NytN6kPpjZW9)!~ zaqKhuQ;qX*7hcBKnBlpt_a|1yrWo>EpFhm6(E8lcf3H2KwhY1 1>AX4|Hu27 zT9QWNX5DMcUGlfcdy@Z5K2Tp5m`^}@;at`H;-dR`M@Lw%hSWW8?Sx?%g=_IBUdNYs zXs`a+UhVR72R~uveKubd8)7FM;;;~lNAMcH!qoe1Lw{m9Y>ESLIxfYXIQ~DKV>N5a zBHV+{Z5DpJY%65LV%P}VVPBk#lk}HTuavAQOK} Iz`gho6CbqgeK?{0InOFex|6z~R>H#P^w+TORFzuT1E=6>JcjS^ zjc&mHR23zyHf*edj_N(BqQu~S^w<6z)A1lWUWGR?=_xy~GU$yC{)OA|JpRNIr|tT_ zI2Kpp+0%Odn*qEak^PKap&^Fi0^E-eFe|&V2v$WebkjZJcC><&!(j5maK=H$9*fl^ zu4qNicNOFm@qPU5kgb;oYhYVE`)k&77IY&&2&dv2+>2Ko7CvBxe{4qu@fHog!i#k9 z5MA{2HK$+lEB5rwg~N42%w&$&*&cV&Q&Q?rt{Zb=E!uY~utH&GcP(sAz9%|31GiyY z>i5C_$Un#5uG)cR#}e4m>PRmVu4nDBZ^_7F*n%#^18_2K#jE%R(;v2bun2nLc>2l1 z9$T0YJ2hoD1)Vuv?r?hCCjSM8&`{QEc8jJE|Am|J5Ei3eIXbLNyp9dKWa>G3(|ExO za>R~29eQ9>R%}3Amj=6%AA!X=wl%Od&ZB%Ap2R?H&oOtcp)@=hqd1na#GCLmK1SE0 zcEEYCHa?+)ubB0^-D3fi_s5Ax_53#jDE+GfTUMXggE E^7>!Htx2ER5UbogC9&(XmSz600)xBnG$x6O7Hp37cgp(Z> z9{#C65cG1ee8!d9klT+|auDZJZv!5}Yxo8eH?srkuWoqQT1Md<^xC3(qG6DPH`jW( zy~K&=t>080?ki()1}?$Pco>_t(7)xZ;3XG{U*f^e`jYCdk6b}l{i=p(H*b6WKf!6e z kzv=`$|y%CO{|}!pRT*GU4XP3 zp+7qFy`7w!pd%0OAonoYIQ{wUjvXW$z9GL~XLsbt_R^j8TqfG}q9^HOzR*E`_sK=1 ze}F8E*1s`zF8j+Sv!EWtN>p&C;WAhq|E1m?e2qWx_o=qMTv!BCGmzq1FWDM@Iia!R zv8#i`>4I?}Bc6bB@V$=wTzm&fGTE99^XWj71qR4UwyXgIFHU(CY>g{vuLJdJMeF%* zM$(N$Q?}3_`{4k}x6{B84A3oc-{~)*I07$GKA!cuFo1jH_cP$k6ZKe|V>*pvz5tWC zX #w-tLB1|n|Q%to#jKj%TauQA2H1lTb>u|*AS@|m{kHW z++iUK*W+1yfZs4vP1`^btc~q43@70d+>A%?E;{d7NL|Y|?1rA$1pP4#BXKEi!xMN9 zrM7K1J-T54*G+72>f zK`eu{u{CzYzBmjg<6PW?r}3`Q(aG?NMB;k3<6Kw ?FJJ5~iB34dm)11+flhTcB4? z56#Ew={T*EL@m<4?cCQ%HemIode)>3l)TGzefL1|#OBNO{5Of-Bt|iSiZoP)hU?)B z%9mqP>iPXDr=8uDAIE1r4}8I-jqLM5x<>l^ZypqKkw~cz9{UGplicWmMJWHgLT{gI z0;T>3S8-c0S6X6w{BxaNH$pqh{k3`tuHRWI64%GJ7{DRv=diF|A1JbpnXP}tVm*xi z;S>CUY2Vs tv{B!y%2h0Vs|QW;mw)(tK1AuWhQ=8Bw~=OIpK z5<^K8)r!gDic4+mfc>a924~|++=bm%$^S`v5Adj}uKjyYW`Iy9p@tqH^dg~yfDCm8 z69^qa2tzU |4Ooe`Hg%AZj#bC%cdb^8n8Npf73O!lE#$?uuIJ$YGhgd*rFW?w&bPw?e`8>IS z+-*?nA-<*H3@H$S3|NnBL-r&`k;_4u;7al(a@EV&5gqLz){)!EugH_+RY?ts_E%JZ zHe@ey4Cy8VHpluPrNdltSfgMK76@lP zm@77rpOL?i)?LcKDcOS@Mcz){N%ms`-}O*mKsilbCG9 km#9KR{u@qQ^y( zBkIY>Z(igq$w{ZnbC~7uHSDtQ@wyIv#%o>){1?p;1?00#=pFJZ_39i|`d@v3M?SUO zOtFvmbbltxqzB}FKJoAm<*}veA=S(DkKU~GR(^&_2X{7KBsY?ul1In{h95wlfn1XT z^UX&q=rNl-@`Jp&tcxF(sP_W@Bh|?iOS(vehw^48IF{hT)=)tMI!W*f-#(89VP+ z_kM?vllIH||G2q4mx4A&Rn_dFu^n+^8pRI}bpquN3ymV5_G2M|9~OF(?8zqD+#|~P zJ(2z7599^1T2JL4M|SF|DZxMrQphwiKrSbrCf_1=k_}kWhH2_EyPnx-68Wunx<@79 zru0ksVgi{?F4TCDJPzM{#g9eIAhRao;NiXk5g?b5Pmt@$9po|c0$F{M3Lj4f*rfU7 zu2FcT6JK1{j`Cu8I;s?(SbdlZZ;&68-;ixj$_2(MF_cw4p8TFo*79d;zw!Og!Q>;f zt)ChG0{OD)WtynEO<8pl@?Cfg`C+O^O2acC3zv(oln*CUDW68Sm%NkmW#ltVxRCx& zkfmr6`Gw?z>#%73RP%`KUaaNt!LhV*+)6n;TD 4 zBe|12Nd88OKUIRY$s}YXzTX`2>n6OSQ6p1KE76qT5(Ss>0KvsChYKqn9FTnYJPsoF zJH^E^^y^3cqTVlx`!A|5q&lYKBXh^_4Y?VZBWf=bcVK8 pkD+<@6PaQS`Rq1? zYmz0Z4#M10d%Sp#@xG}2js&VW#b)wj@@KLc1Bf$9uNJuvJUYNB2GG9^)&Ef8n;6ubq6Y;|ataw@0{>u1HS5e=5xZNlE!m%(K<1Fk$@SzH ))rBELU^{lD~h`xv^-9G}=p{!CsX zYyY79+mW}BQ_1DzI&wdGimY~=;mMKYtmEkaGGUE^mE_yxUb2j|{;2d@kPc*U#W 2>ZB z^QvI^y?TN8HV*F3?h@_0B8P{@itjgIVEHaXcuR0PHo2b|(FspT-*vZ`M5dELvXFd? ze2ILUTt67IBMu}R%lTiPdJV!Zch>5;Vgxyfyq(M-b4kyKxP`Lw4v|AHpj@MS8QrVM zhj2W%@BP){3GxMU**UELoh7S9i}NV)x2we*aBb96q9fhwsJM~bO8ThRlX{Ej9!U3a zGMU`Xa0kdE NZ<8xrA)PJ0=cp z!zYl-81G+SVX=!Jy|rPt4dVBIcz|#_86ZF6w)qhG8)?l``~T)-Z_-6hC#y|Tw`%U7 zdnvh+TxU?5CU#PAkaY7v;a9pZlMQgEQ=V|NCkK+HGqLK$4;L+G#-CwHOL(5}H{|jh zVIG(DzrCo=5$;)o({c^R<^7*SKgU7BNb=^RY8SkQ%jP%8iA*qyOrhQ^ve9OB9#EUh z?gX+E bTgv) Kx(lS4rP`6Su@m^x7CV+JH!4aHo60|jlrI#76w{13U646 BAZN#n&n z@<(zyOWG$5V>^Dg$3I&gD2$-HnCAvhkgt+k$?S3JJYg}}f08;67*0+mwJcsNAs-{( zB)5@YlVzlstRj4KyV~P@cbod$b}8L^$y4OTQR=`U)~VQ0Q4=F6Sd*>}AgX<+B1|BM zkYmY#C)EMOGi;(a$nI?86dpX}kw<8ZGZ;RH{0|1QwKcTc#9|6oknh1`i2XM4WSTmF zc!Bcw$=500NB0SGE9L(~{=$SW)8CG5yX m7R*@OWIU&FZcDl&|>Y~C{8|GASR*I%+t&?Ntm zHP I6gvmpDsS6{H;^?jMvMnL>wobCWkRQ-v7Ax14b#IFR~p} zoPmRg()~}0crbd!lcKA}i$P=t6&8?% -@!p$v&huj2Dy0 zIphMcL*Lb6EF#F<_M`~W9a@fo=<-bQPc}|8@gouotD$6@7l;*mQ8SHJ; i^V(Otoe&YpCon1b(o?ik|~%%W{`ez4Y%Q6va0W8Q?+2zv?qI$N4mi{ z@ssVD7m$IxpeO{7z8n;nU&neN*CQr$!Lt3x-J;z+&^wOrf8A_c6pL}lq5NH6`x{EJ zMOSn~{N}6Y9rRZ`GUer#D)w8oGwJ}w{FE#DlPRQ!TtK>8qlWOay$j+o@{M04F89Uc zg)gJp`+*uRy2=YMiF}tQ7vGYW0qP{96X_&9WD)r~naX$q 7Na{suNU>Ek@u3XGQy?QdybsXrYj_CoxxHp?FrG6?0E+3e<^TM zFoO(`MdVZDo8(92A@U4a?JN@@`;cSEJzQl^8;Rbt 3$Pv8lN?2M_QC&y z*5YO|g`7doC5y;s$alzX B)o?YnH&b=y2gnUWS56< z-~X$wqCYu;yo9G_ pHrD-v2z+*qu>7taX{>LlJGx03tG6XYebe!kN0n2+^`40sC#Gsryh6DBZ+ zZ8j?vv)B0MLL;9dw~~j*zsNeXlzw}1BsrJ7U*p9F@&Nf8S=bh*-#K>iGHG|K2)mIk zaxVD@`3`xEtTkJO>rH;l__QPnPEyd_Aw PUZYFdi+VoWvU3C zqIFx*N7v8SO)kA@j)l z$ydpb$YbO`WRngmo_$&9|1tw-%mnUGIqgBZ$;IT; @(T4@&Q p@%{13Sc zTsx|vI6|H$wNAGvkI7^Xc|Z9A>B_;gTsiR~n_R-0YFtf~@D_45Sx7!fmXM#4KPqad z2NgjRasWAr^pPt_>rFz4r|XN?=-y5KLRPP?;%i9`BWIF6@&SWdRq-DRc9P$c)oQ4S zdXOn(kbIPUgZzX%PF^NsYpQrUkmJapqz2=7f`V!-F bF{jd+z#Cs30K4x(TjnNBVs@9Kp4ykTAO99c^KKwcpmbyj-4$Ye5y zEGFL~KkAJBFC+e%g1^Z+T~tI}NEew--c1Vp&5E`O(M*W=9koR)*@=9LMv7i%=*?&M zYY6$Ouj0i`d78RdMvv#njpV1~Pvk|iZUfbydy==3Gs&*E;cj@askoh7PQFDRBroFp z$xw`&rdYrTo+V4jkICa?R6~`42IN3;Cb^h=k}M^^A _%+M0pxhnONPkT$Q|Uj zB#sjdTZpwtYBtfyxUdNq`9d3fRHC_E=rpqT$m3-7uF5!XB4?4yNgVGP34Cx1GK?X2 z5#5$M;dZqaEy-b$8U(2nEG8c(-y}aGzbF4DA7R9ko8!sp)^+epDyru4$>-^h&99N6 zPf6T+Hr(H!Mm0G*{mH7mOpoGt7~_^YqDGWbY)-m4klaM~Eo7U;7*Ni&7yZa&(nH=u zt|r%$I~Jq=%Ya8HxI(tMTSeHPoIu`5t|Hfwd&tdW5N>Z{ag4N%RqoE@Sn^Ktg|X=W zGQchhekEg4l)M)?om@z+BG;2U$)CumaVmU6vKKj?%p(glUaThHB=?i&$qB6L9EVz- z7m)AMSUcSWv&O8LMLnFf7;XGG`3Ctk!{MOS@Rz?oZd_a?f2Gw{AEK-`fqZ!`+>_(Q zcJe3EmaXKi$U)?7WPqGU+j)e1OHmVhDL6$|$7CXRJncy9JZwU%C5VS+VlQ}0bMYei z9=VtNnY5%Ty~bpBax__-gCV~t9uGT8YVl$R1wWCNT&2*097<-8i^xaGH^}|uDbkXs z;`w4Ax+T_nXUO*;mut*^a>&;(Nq9K+7m;s}Uy%QhU8bu!VhlN-e2y$7e<7RPuJmsv zXOo5GT2Pa-)W;P3O4gX60w$6($i?LIWK*lMo$lmlvU4ona(}wIm_>IUnQ!XRKRp)x zUn(BV$1?lbPU3fR6%}76OUW und2YEla zj%>k<^&!WR^T=Xy9eI@D4$ys;tUo~&P`iZ}bI2m{MREt3c^B6IqOE8(Pj$z!WG-1u z{+HZMK80}Jb?wCm@QVN~P(PV&pjNCweLtY_IGND@YRmQrJ$z(410Qm~J z!=RQR$|#6kpd#o;P9c5dW8_=pSL7A4<3bg_040+(l}xvne1a??4@znn+5VycPv96M zU3bz&eng`@Or9pI)ltUVg`7%aMX!1XxsCjVtZ7&Noj^@?qY)I$AnzfcB}>S!$X`fX zjEblw>9|QvLVf8TN9K_Cl6x?ItbezuaL+{lm)q=Oa4hkYZX<}0+wWbJSHbBPPdas$ctq4R?5E>*@twJ>EuF#S{>nImEKQ} zC&~5Xhvbjsd9q$>l~8kLuruAal4;~3@=3CU+$X7_B-5whUCbd(#c%Ye17njVY)keh zZ(%|c$uX3>hv2ci|27l_ E}= z8_6%s43g1PRfd1PTZkdGJBZroj*{KT=JfAJP9W!yzp(rMJ{A3cNmf&Vr!5R>RCtU` zg5uGn4q`3cyU4?2eU{9}8d*wqrToi0wbWWo`MYE~Yi@59_W$Gb_=jwQb2~XMbcBCO zw~k^M=_VJEcz(_Be~bKryhJu!qWt@i1DUa@ 3Qx^>l}^Cr?s&LIoQ z)#L{Twd!I&1wWCO$k=)+!uQ!-4w6^Mrrm|H<+_D*llPJ9$-U&CWZgs+uD7IyJ=<** zEG7R#9w0A}O?s$+gUKxNNpds!Gg+sn(jP$1B$tuvK}~KNzowvSFBPCWIg`AXe2M&+ zJWV#~t->Xf^T-#-kIBEtmVK1|us%p%X246qqvX5fw`7gJ%ov$M=8}(-?~{kf-t|-y zP9Y!9R-4rg sJNW|{b+d|~6*-ig zPA(=NCD)PPkQYe%Kow7C(xHgo|D_ RY$9c8j)Se zn@JZrh0G)GAs-^2C)bl($ det?2(U12MO%3^XS@kweMxWCpp2TuDApzD@2V zj|9;FA88~0BBO%JU59KMME{oxohTSUjwC0L>EwJepL~@3CRsg*5IYpp?X7WQ;6x#I z|J+1enShnj_f14ixFwsDUC7CrAM-ID!N?%zlgr3w$o1q_@(c0=d4a4QQ1NvkN0O7s z3^GV+5Afn?aviya>;xl`Pp#ZcMvuYb63=$jBioR@$&uu3WG1 q5&0l;yit=)h&*Ye~>jFRq_OK5a|I=wsDL3WFh$^xt`ob9wNtqGM! NcV^2Ve%JpFT=GeN4NWFwCF<~WP-nt-|)G=tEQacEE*z`B@q9p zjMSyOBUzL3W@K+NiHt@>vcy?*FC|AZyiWIX 3iZ_8aRGuaG}b|1WYpYoOLuRiN$Eo5>7)MfY!tnrKf&af^B>tVKK?Fvh3X`cU2x ziZX*YlXmL$rrt0MUQ&y`Rg9qfHoE;}G5G@dCRu7wyH$)~TYp23vt;!s6>)2_JLw{4 zl76y?e4d=n47Fp1hmkYMm{E9sSbJ=ONM{BXkc%1cA+n5*Qol_12js)1k(y+}@3zKU zF5iz4Ls^4k$g{NinRM?%JesWPAO#PT|0TDOqKkT#`)j(}%A6zP`-h#Y-bn4`zSv`cQK!c*0dy? z`{M&-50FoiuaIw(Yd8mNq5Dm`_t5<%X+igs16nZ)@G|*68B70$qw%g-;zXf!;KeX< z2ANBa!~nDQO_x|g_cP>n@(_8MY<#PVxI1~4C2l-Kb?;%6PgB%H$_RYf$U8wa=1AtI zdoOc*l)OOJa;heaBYQC1B8FQ >CV2eaxXa(d3@JinZr zM*mmn4$|G3;giWM@;KrV15?BmP?Hypx+y^)aw_R1?;~F(_mIcPf5^5ssdxsFsia2g zWC?lbCiH)qfeRGayQ>I#kW kJ6>ZSA&$f2Z%ypLQ%mXb$FYi||49eJz93m^F*+x{zZ4@>qv zd5-*r{+TSvGS Yby(aV+iCd%K~X*e;4%Fdu_u(<^)oS8HRHx*b8ff)aUdLCOxVj!>-?qa zy@5C%j^CeeEiyYnAJ3H3!Eht*CvHT1@;2Wc_dp ~R3%<4EC%Jr h<0-$L@o4**lOM>lWE39mk~xhf3lLHC-@2H92ZrSLzxLqu+Ka7LiTB`M zGhmh2&I}yoru5ri)l00^Pvh0DeXB&H6uedY)g;l0?5~LP9|}^*ohuRXxD}%I35#&X z;wu+qZ*n9#h0G!EAy<+wlO^Qk9e9lopK7f4k>cAM H(G1;2zL=GfN&qs+=UwZ0Xcg9?IYFc*CAMp6z ze(y>bOO>?jJiq7B*ni@peel3_`rG^enP$;vyDbghZeQ-R>e-E}HYr&ZWy!2=cV}nk z%}<^0p6dyEQ{91*Gc_$qRcd;2-LtbjseZRV8xhsl2gF)(qvy^`^?S3k0{X1^v298= z#ajH)C8L^H7F4O9;R&Q=rsg36uQ%1}^LX7pkA9{{i rscEq~ z(z5l$)|PtufSl%a_37T&UZ!ZCyeQf4jeW>k(zKoB^F~_Q=m}%Tq&m{m9X_9X;l#YL zv+wYv1w?9CJ~S^c`v&5X?rgv3269)xlQZ6v;qiHL)8KE!8WGzy-ea|ElamTJ4)!mJ znc`s+B9a*s%)U{k#{_d`dwgaHsjsp0s@Rb}^ce4+e*@`+fG;a|&J8>!`hwTh%CNi~ z*;Z!uhWR}1fQJ>s`j3!>E3dM?Y=+8`;dw!rPo)Z0(HKorS$AZ1o;z>@RhO1~BeAT$ z$`PoVt1PQ*Dk^Oy` H^{82^nZz4R=gR{a8I@;aE(ja)kt4n zPNjH XGK5tPJ=vXIQJs*g z%v}>@h46;CMljW4l_GPx{T1OvD(&Bun-y^8<<0e*Id F25^xo;y1$Jshh)(z0D}X>vfQ@K#x?a7S3I%dbjP;TdUi$yvEs84FFn6t@pw z6vyx)2NA 1dA=NXAVNOgqj|I4X+{UQrk=-pB77=W%MHA4XaEs<*W2v`Won~5KBp(! zThRuP8}{NiMp(15CVPDTth`*)nc?$ Z8 zU+lWL#`^j<;;QNoz7dyXidyQ2-iUk7bTusru8Z4aDY zwL@C=(5&2aWNKtqHac! +@#h77X`WQCG)UjS#xxRb3Ek_{8Ea-T_4*Q~gflo1`I8E|yKjreCC^N-oU*n@ zNzl>4nc~z7p-d>49Y}Q>3VLEr^Tzs~ewLd0GgB;cWkW;;Q?r8 f{!UjEndgmNq4!$t{X4de<42 z>UzzYmbz8bjJ3_XGcDiPYNFMAc?(lzU)4XKZMn+|!!H? di{m9VM+B=<(T> zDkUkK+XgHpOR_DCs#-Gx0X;U}QeVS`H+7CjS+(g%y~8gD0aTNcWarIxXQyK5jZni< z3vEzY(Tnbp>RI5$%Jv!`pXbihbd}SJcp=ne+No)I>B_u~ 6>r z7fo90gF1JJDe2(1q}9}h1%276?rd*vR(cQAvq+^N;GdU<{fNh%8}wcy4EWF&Bh^%H zBPCcRV$P^6$(ws^OsHW^PH&NVez#l?84_eaFMA#ax->Vo0A`vM(n^^Ni_?SNOm}X2 z_I0tK#@*@Z7)>(rDiY9C3h?virOmC7Q>^Tway5igniFX?m2(_XBVqL{ql3$c&8mSH z)1R9q*U6RRnD5RC$RPnC=;encrRVx#rx@unFx_4wmy vENWiTdAyq8odVbrnW4a;J(CQBM^K`8+v!^E@{$qR$fu zV#bro*X6)(tk8`@NvoHA8`FcSCeIn9!F#mGyr-tS18$5o#v;R1MdVodhC8K^v%Fl4 z$fm(c3947E9Ky9IyIHDXkP#6XLRliY%TVo)p2r9xIwrabh2g$}@z0;-H*_NujWU=; z2xR48zlNt=RMbH)iUV8D&5Z~^doYGwX}y|UPnl*hJD8Q7j V%IR1PY&G^k0b4;i~&7ZZ@*%pBvCvtTwVqOvk-@p@q; zdot3$UPqH7bw*aU$4n*zXH%J|eW`EC%waQC4nW>KpP9JKi|NPwL=3W+Cn?PI^l)eK zI?LNb2BFKZ1$>ci%xeoG-C2Hjc5V>cPs3QG3xj*ul?F33Yc0#4o;62K-_)Grb9>px zQnUOxHIq97`pG3FH8ptpy%j@gmOto)sfAS+M8=AKo1WMwb+&1|bccrr>F(JhHS9)Z zq-JN~yJIS0IAteM%S O=r@I6xoCl}+laSjiw>bJ 5( dtD2$0V6L!%Uw7@L`JBpOnQ=t+i-toc-g%)#CutOz}C zR(LWp^e{X50;w`t!}Uy(kGU`^-^O z-Oi9&X3TPvh9NM^TQPOWfmemB)QEYSDfaqcbc{geTQ0PX0H(}4uh0BEpFCJJGA!p! z8L8Ku6<$CiIcYNJ8<@ALmxry(wS%!S{}{Jy+?rCSa;4S2a-?FB8*X-rP$}2#R#N5a zCl8PIs_aoSn2WL0T)`+A_ES9nk>$Yr=jF+!QLIdVxSPw%ia|8&TghCdimZvW*-Y06 zWUeV9WyWqgQlK_^nh`H54WooVke8E6Bc!JEkOjjkE*;%d?%7oU<35->h>jP? MzBYw_suAs8yHHJeBM?Ezi4 |kUNyNF2H?7Vaw@2V&Q0nM8#1rhl*&H!ZQuqaE#9s~#*82ya(c10ZCYnE$< zR(f98oDFvd%rp?WMm40Fn$hPY {o`KT;dq{FzQ*mrf{&)pgd &KfKv5_&tW7G=-{SnJ5LyDo$80IU&VZ!#g=Tr=# zsLEWwx}8|Dl?~$jK^*}lrY3r^^yMfghd^Z*m^#&hEltiumF3~VJ}ontJJ*wr{emx| zZy6CAUX}U`3_}TJO_<}KQ3%6axBy7S!n0y$sT!e@i1l4b#!Ztk+^aiRMtzv#2%9iP zTI2#tZRF&7C}MaqHS==kV7Y1TAic6f4I4XV%!pwV*; |zX%zGQXvaD8dQmki zqeqtND4IO4nAa!OkDg~Xd$!!BAfjM+`csn1YhWddm|{$Uyx)Vf@$l+J$<;oZ>F1>P z;#L9a4|k0kF=1kA^6=jDHRh@cC$1_@ZPFc)taMhI*mBZ)hxZ{l7`u!kib|u5AvN0B zt6>T^GbMFVeR6NAj;W2tGnBD##Q<7KQ)b~>8E@uaGl&{7NMnmk$mf}lGYWjnvLX?= zyYmNU%Sg=u%?KLN0}XdDH*z^_h%ip8ycoU25Jjxe!fu>JvV$2S?6WoWMa WP+0s+CS>HKms0ULbRh zO2*?JutJQZ?wqg@8Zk)~IX6X=@Op(nUH~Vc9D^bb{ESq**`5V>b_N|R+!KNtGLM-; zp7exIZ_tY>k7BZ8Twie2>5b)yx$;yZ^H55eOt^Z?az~2&8iqwIY9eGD|0~Hd(kr%g zs>0+>1
I; zV4gU@F&lVJj!eolKa8#BB4+;Fbho_AZluqkW=eB>{^1p?Coh~ySvgsODahZ%irA!l zygb;L$c2S`Cc?Bx={dm@7@jvD3vhS(FgcfnPn9A)HTcSa5osgbB=5h5QeM{vkvqXL zdAJ=sOxCAN;F_FHG7dxOFZYYtvPgJr!U%v8m^CIZF@Kmxcj%7BeP0!?VL8Sn8#Z!t zV%)C?ciOb