Confluent.Kafka
2.6.1-RC2
See the version list below for details.
dotnet add package Confluent.Kafka --version 2.6.1-RC2
NuGet\Install-Package Confluent.Kafka -Version 2.6.1-RC2
<PackageReference Include="Confluent.Kafka" Version="2.6.1-RC2" />
paket add Confluent.Kafka --version 2.6.1-RC2
#r "nuget: Confluent.Kafka, 2.6.1-RC2"
// Install Confluent.Kafka as a Cake Addin #addin nuget:?package=Confluent.Kafka&version=2.6.1-RC2&prerelease // Install Confluent.Kafka as a Cake Tool #tool nuget:?package=Confluent.Kafka&version=2.6.1-RC2&prerelease
Confluent's .NET Client for Apache Kafka<sup>TM</sup>
confluent-kafka-dotnet is Confluent's .NET client for Apache Kafka and the Confluent Platform.
Features:
High performance - confluent-kafka-dotnet is a lightweight wrapper around librdkafka, a finely tuned C client.
Reliability - There are a lot of details to get right when writing an Apache Kafka client. We get them right in one place (librdkafka) and leverage this work across all of our clients (also confluent-kafka-python and confluent-kafka-go).
Supported - Commercial support is offered by Confluent.
Future proof - Confluent, founded by the original creator/co-creator of Kafka, is building a streaming platform with Apache Kafka at its core. It's high priority for us that client features keep pace with core Apache Kafka and components of the Confluent Platform.
confluent-kafka-dotnet is derived from Andreas Heider's rdkafka-dotnet. We're fans of his work and were very happy to have been able to leverage rdkafka-dotnet as the basis of this client. Thanks Andreas!
Referencing
confluent-kafka-dotnet is distributed via NuGet. We provide the following packages:
- Confluent.Kafka [netstandard2.0, net462, net6.0, net8.0] - The core client library.
- Confluent.SchemaRegistry.Serdes.Avro [netstandard2.0, net6.0, net8.0] - Provides a serializer and deserializer for working with Avro serialized data with Confluent Schema Registry integration.
- Confluent.SchemaRegistry.Serdes.Protobuf [netstandard2.0, net6.0, net8.0] - Provides a serializer and deserializer for working with Protobuf serialized data with Confluent Schema Registry integration.
- Confluent.SchemaRegistry.Serdes.Json [netstandard2.0, net6.0, net8.0] - Provides a serializer and deserializer for working with Json serialized data with Confluent Schema Registry integration.
- Confluent.SchemaRegistry [netstandard2.0, net6.0, net8.0] - Confluent Schema Registry client (a dependency of the Confluent.SchemaRegistry.Serdes packages).
- Confluent.SchemaRegistry.Encryption [net6.0, net8.0] - Confluent Schema Registry client-side field-level encryption client (a dependency of the other Confluent.SchemaRegistry.Encryption.* packages).
- Confluent.SchemaRegistry.Encryption.Aws [net6.0, net8.0] - Confluent Schema Registry client-side field-level encryption client for AWS KMS.
- Confluent.SchemaRegistry.Encryption.Azure [net6.0, net8.0] - Confluent Schema Registry client-side field-level encryption client for Azure Key Vault.
- Confluent.SchemaRegistry.Encryption.Gcp [net6.0, net8.0] - Confluent Schema Registry client-side field-level encryption client for Google Cloud KMS.
- Confluent.SchemaRegistry.Encryption.HcVault [net6.0, net8.0] - Confluent Schema Registry client-side field-level encryption client for Hashicorp Vault.
- Confluent.SchemaRegistry.Rules [net6.0, net8.0] - Confluent Schema Registry client-side support for data quality rules (via the Common Expression Language) and schema migration rules (via JSONata).
To install Confluent.Kafka from within Visual Studio, search for Confluent.Kafka in the NuGet Package Manager UI, or run the following command in the Package Manager Console:
Install-Package Confluent.Kafka -Version 2.6.0
To add a reference to a dotnet core project, execute the following at the command line:
dotnet add package -v 2.6.0 Confluent.Kafka
Note: Confluent.Kafka
depends on the librdkafka.redist
package which provides a number of different builds of librdkafka
that are compatible with common platforms. If you are on one of these platforms this will all work seamlessly (and you don't need to explicitly reference librdkafka.redist
). If you are on a different platform, you may need to build librdkafka manually (or acquire it via other means) and load it using the Library.Load method.
Branch builds
Nuget packages corresponding to all commits to release branches are available from the following nuget package source (Note: this is not a web URL - you should specify it in the nuget package manager): https://ci.appveyor.com/nuget/confluent-kafka-dotnet. The version suffix of these nuget packages matches the appveyor build number. You can see which commit a particular build number corresponds to by looking at the AppVeyor build history
Usage
For a step-by-step guide and code samples, see Getting Started with Apache Kafka and .NET on Confluent Developer.
You can also take the free self-paced training course Apache Kafka for .NET Developers on Confluent Developer.
Take a look in the examples directory and at the integration tests for further examples.
For an overview of configuration properties, refer to the librdkafka documentation.
Basic Producer Examples
You should use the ProduceAsync
method if you would like to wait for the result of your produce
requests before proceeding. You might typically want to do this in highly concurrent scenarios,
for example in the context of handling web requests. Behind the scenes, the client will manage
optimizing communication with the Kafka brokers for you, batching requests as appropriate.
using System;
using System.Threading.Tasks;
using Confluent.Kafka;
class Program
{
public static async Task Main(string[] args)
{
var config = new ProducerConfig { BootstrapServers = "localhost:9092" };
// If serializers are not specified, default serializers from
// `Confluent.Kafka.Serializers` will be automatically used where
// available. Note: by default strings are encoded as UTF8.
using (var p = new ProducerBuilder<Null, string>(config).Build())
{
try
{
var dr = await p.ProduceAsync("test-topic", new Message<Null, string> { Value = "test" });
Console.WriteLine($"Delivered '{dr.Value}' to '{dr.TopicPartitionOffset}'");
}
catch (ProduceException<Null, string> e)
{
Console.WriteLine($"Delivery failed: {e.Error.Reason}");
}
}
}
}
Note that a server round-trip is slow (3ms at a minimum; actual latency depends on many factors).
In highly concurrent scenarios you will achieve high overall throughput out of the producer using
the above approach, but there will be a delay on each await
call. In stream processing
applications, where you would like to process many messages in rapid succession, you would typically
use the Produce
method instead:
using System;
using Confluent.Kafka;
class Program
{
public static void Main(string[] args)
{
var conf = new ProducerConfig { BootstrapServers = "localhost:9092" };
Action<DeliveryReport<Null, string>> handler = r =>
Console.WriteLine(!r.Error.IsError
? $"Delivered message to {r.TopicPartitionOffset}"
: $"Delivery Error: {r.Error.Reason}");
using (var p = new ProducerBuilder<Null, string>(conf).Build())
{
for (int i = 0; i < 100; ++i)
{
p.Produce("my-topic", new Message<Null, string> { Value = i.ToString() }, handler);
}
// wait for up to 10 seconds for any inflight messages to be delivered.
p.Flush(TimeSpan.FromSeconds(10));
}
}
}
Basic Consumer Example
using System;
using System.Threading;
using Confluent.Kafka;
class Program
{
public static void Main(string[] args)
{
var conf = new ConsumerConfig
{
GroupId = "test-consumer-group",
BootstrapServers = "localhost:9092",
// Note: The AutoOffsetReset property determines the start offset in the event
// there are not yet any committed offsets for the consumer group for the
// topic/partitions of interest. By default, offsets are committed
// automatically, so in this example, consumption will only start from the
// earliest message in the topic 'my-topic' the first time you run the program.
AutoOffsetReset = AutoOffsetReset.Earliest
};
using (var c = new ConsumerBuilder<Ignore, string>(conf).Build())
{
c.Subscribe("my-topic");
CancellationTokenSource cts = new CancellationTokenSource();
Console.CancelKeyPress += (_, e) => {
// Prevent the process from terminating.
e.Cancel = true;
cts.Cancel();
};
try
{
while (true)
{
try
{
var cr = c.Consume(cts.Token);
Console.WriteLine($"Consumed message '{cr.Value}' at: '{cr.TopicPartitionOffset}'.");
}
catch (ConsumeException e)
{
Console.WriteLine($"Error occured: {e.Error.Reason}");
}
}
}
catch (OperationCanceledException)
{
// Ensure the consumer leaves the group cleanly and final offsets are committed.
c.Close();
}
}
}
}
IHostedService and Web Application Integration
The Web example demonstrates how to integrate
Apache Kafka with a web application, including how to implement IHostedService
to realize a long running consumer poll loop, how to
register a producer as a singleton service, and how to bind configuration from an injected IConfiguration
instance.
Exactly Once Processing
The .NET Client has full support for transactions and idempotent message production, allowing you to write horizontally scalable stream processing applications with exactly once semantics. The ExactlyOnce example demonstrates this capability by way of an implementation of the classic "word count" problem, also demonstrating how to use the FASTER Key/Value store (similar to RocksDb) to materialize working state that may be larger than available memory, and incremental rebalancing to avoid stop-the-world rebalancing operations and unnecessary reloading of state when you add or remove processing nodes.
Schema Registry Integration
The three "Serdes" packages provide serializers and deserializers for Avro, Protobuf and JSON with Confluent Schema Registry integration. The Confluent.SchemaRegistry
nuget package provides a client for interfacing with
Schema Registry's REST API.
Note: All three serialization formats are supported across Confluent Platform. They each make different tradeoffs, and you should use the one that best matches to your requirements. Avro is well suited to the streaming data use-case, but the quality and maturity of the non-Java implementations lags that of Java - this is an important consideration. Protobuf and JSON both have great support in .NET.
Error Handling
Errors delivered to a client's error handler should be considered informational except when the IsFatal
flag
is set to true
, indicating that the client is in an un-recoverable state. Currently, this can only happen on
the producer, and only when enable.idempotence
has been set to true
. In all other scenarios, clients will
attempt to recover from all errors automatically.
Although calling most methods on the clients will result in a fatal error if the client is in an un-recoverable state, you should generally only need to explicitly check for fatal errors in your error handler, and handle this scenario there.
Producer
When using Produce
, to determine whether a particular message has been successfully delivered to a cluster,
check the Error
field of the DeliveryReport
during the delivery handler callback.
When using ProduceAsync
, any delivery result other than NoError
will cause the returned Task
to be in the
faulted state, with the Task.Exception
field set to a ProduceException
containing information about the message
and error via the DeliveryResult
and Error
fields. Note: if you await
the call, this means a ProduceException
will be thrown.
Consumer
All Consume
errors will result in a ConsumeException
with further information about the error and context
available via the Error
and ConsumeResult
fields.
3rd Party
There are numerous libraries that expand on the capabilities provided by Confluent.Kafka, or use Confluent.Kafka to integrate with Kafka. For more information, refer to the 3rd Party Libraries page.
Confluent Cloud
For a step-by-step guide on using the .NET client with Confluent Cloud see Getting Started with Apache Kafka and .NET on Confluent Developer.
You can also refer to the Confluent Cloud example which demonstrates how to configure the .NET client for use with Confluent Cloud.
Developer Notes
Instructions on building and testing confluent-kafka-dotnet can be found here.
Copyright (c) 2016-2019 Confluent Inc. 2015-2016 Andreas Heider
KAFKA is a registered trademark of The Apache Software Foundation and has been licensed for use by confluent-kafka-dotnet. confluent-kafka-dotnet has no affiliation with and is not endorsed by The Apache Software Foundation.
Product | Versions Compatible and additional computed target framework versions. |
---|---|
.NET | net5.0 was computed. net5.0-windows was computed. net6.0 is compatible. net6.0-android was computed. net6.0-ios was computed. net6.0-maccatalyst was computed. net6.0-macos was computed. net6.0-tvos was computed. net6.0-windows was computed. net7.0 was computed. net7.0-android was computed. net7.0-ios was computed. net7.0-maccatalyst was computed. net7.0-macos was computed. net7.0-tvos was computed. net7.0-windows was computed. net8.0 is compatible. net8.0-android was computed. net8.0-browser was computed. net8.0-ios was computed. net8.0-maccatalyst was computed. net8.0-macos was computed. net8.0-tvos was computed. net8.0-windows was computed. |
.NET Core | netcoreapp2.0 was computed. netcoreapp2.1 was computed. netcoreapp2.2 was computed. netcoreapp3.0 was computed. netcoreapp3.1 was computed. |
.NET Standard | netstandard2.0 is compatible. netstandard2.1 was computed. |
.NET Framework | net461 was computed. net462 is compatible. net463 was computed. net47 was computed. net471 was computed. net472 was computed. net48 was computed. net481 was computed. |
MonoAndroid | monoandroid was computed. |
MonoMac | monomac was computed. |
MonoTouch | monotouch was computed. |
Tizen | tizen40 was computed. tizen60 was computed. |
Xamarin.iOS | xamarinios was computed. |
Xamarin.Mac | xamarinmac was computed. |
Xamarin.TVOS | xamarintvos was computed. |
Xamarin.WatchOS | xamarinwatchos was computed. |
-
.NETFramework 4.6.2
- librdkafka.redist (>= 2.6.1-RC2)
- System.Memory (>= 4.5.0)
-
.NETStandard 2.0
- librdkafka.redist (>= 2.6.1-RC2)
- System.Memory (>= 4.5.0)
-
net6.0
- librdkafka.redist (>= 2.6.1-RC2)
-
net8.0
- librdkafka.redist (>= 2.6.1-RC2)
NuGet packages (691)
Showing the top 5 NuGet packages that depend on Confluent.Kafka:
Package | Downloads |
---|---|
Confluent.SchemaRegistry
A .NET Client for Confluent Schema Registry |
|
Confluent.SchemaRegistry.Serdes.Avro
Provides an Avro Serializer and Deserializer for use with Confluent.Kafka with Confluent Schema Registry integration |
|
AspNetCore.HealthChecks.Kafka
HealthChecks.Kafka is the health check package for Kafka. |
|
Confluent.SchemaRegistry.Serdes.Protobuf
Provides a Protobuf Serializer and Deserializer for use with Confluent.Kafka with Confluent Schema Registry integration |
|
Confluent.SchemaRegistry.Serdes.Json
Provides a JSON Serializer and Deserializer for use with Confluent.Kafka with Confluent Schema Registry integration |
GitHub repositories (73)
Showing the top 5 popular GitHub repositories that depend on Confluent.Kafka:
Repository | Stars |
---|---|
abpframework/abp
Open-source web application framework for ASP.NET Core! Offers an opinionated architecture to build enterprise software solutions with best practices on top of the .NET. Provides the fundamental infrastructure, cross-cutting-concern implementations, startup templates, application modules, UI themes, tooling and documentation.
|
|
MassTransit/MassTransit
Distributed Application Framework for .NET
|
|
dotnetcore/CAP
Distributed transaction solution in micro-service base on eventually consistency, also an eventbus with Outbox pattern
|
|
anjoy8/Blog.Core
💖 ASP.NET Core 8.0 全家桶教程,前后端分离后端接口,vue教程姊妹篇,官方文档:
|
|
dotnetcore/FreeSql
🦄 .NET aot orm, C# orm, VB.NET orm, Mysql orm, Postgresql orm, SqlServer orm, Oracle orm, Sqlite orm, Firebird orm, 达梦 orm, 人大金仓 orm, 神通 orm, 翰高 orm, 南大通用 orm, 虚谷 orm, 国产 orm, Clickhouse orm, DuckDB orm, TDengine orm, QuestDB orm, MsAccess orm.
|
Version | Downloads | Last updated |
---|---|---|
2.6.1 | 24,798 | 11/18/2024 |
2.6.1-RC2 | 274 | 11/18/2024 |
2.6.0 | 422,569 | 10/11/2024 |
2.5.3 | 884,489 | 9/3/2024 |
2.5.3-RC1 | 488 | 9/2/2024 |
2.5.2 | 695,442 | 8/7/2024 |
2.5.1 | 138,227 | 8/1/2024 |
2.5.0 | 760,494 | 7/11/2024 |
2.5.0-RC5 | 428 | 7/10/2024 |
2.5.0-RC4 | 391 | 7/10/2024 |
2.5.0-RC3 | 320 | 7/10/2024 |
2.4.0 | 2,663,214 | 5/7/2024 |
2.4.0-RC2 | 351 | 5/7/2024 |
2.3.0 | 9,333,272 | 10/25/2023 |
2.3.0-RC4 | 4,938 | 10/25/2023 |
2.3.0-RC3 | 3,523 | 10/23/2023 |
2.2.0 | 6,164,555 | 7/12/2023 |
2.2.0-RC2 | 3,401 | 7/12/2023 |
2.1.1 | 4,984,653 | 5/4/2023 |
2.1.1-RC1 | 4,074 | 4/26/2023 |
2.1.0 | 1,779,985 | 4/6/2023 |
2.1.0-RC3 | 3,264 | 4/5/2023 |
2.0.2 | 5,330,519 | 1/23/2023 |
1.9.4-RC1 | 154,695 | 9/21/2022 |
1.9.3 | 12,603,348 | 9/12/2022 |
1.9.2 | 4,623,145 | 8/2/2022 |
1.9.0 | 4,363,140 | 6/16/2022 |
1.8.2 | 13,584,050 | 10/19/2021 |
1.8.1 | 1,139,334 | 9/20/2021 |
1.8.0 | 327,611 | 9/16/2021 |
1.7.0 | 6,953,464 | 5/19/2021 |
1.7.0-RC4 | 9,863 | 4/23/2021 |
1.6.3 | 7,628,549 | 3/19/2021 |
1.6.2 | 1,422,858 | 2/25/2021 |
1.5.3 | 3,195,004 | 12/9/2020 |
1.5.2 | 3,950,333 | 10/20/2020 |
1.5.1 | 2,315,166 | 8/20/2020 |
1.5.0 | 1,721,237 | 7/21/2020 |
1.4.4 | 2,126,191 | 6/29/2020 |
1.4.3 | 2,453,674 | 5/19/2020 |
1.4.2 | 999,962 | 5/6/2020 |
1.4.0 | 1,824,795 | 4/3/2020 |
1.3.0 | 5,074,610 | 12/4/2019 |
1.2.2 | 708,793 | 11/14/2019 |
1.2.1 | 1,015,011 | 10/9/2019 |
1.2.0 | 840,726 | 9/19/2019 |
1.1.0 | 1,714,259 | 6/27/2019 |
1.0.1.1 | 1,121,567 | 6/3/2019 |
1.0.1 | 194,601 | 5/28/2019 |
1.0.0 | 552,833 | 4/25/2019 |
1.0.0-beta3 | 123,236 | 2/7/2019 |
1.0.0-beta2 | 658,146 | 10/15/2018 |
1.0.0-beta | 96,451 | 9/24/2018 |
0.11.6 | 1,785,629 | 10/19/2018 |
0.11.5 | 621,739 | 7/19/2018 |
0.11.4 | 850,522 | 3/29/2018 |
0.11.3 | 575,407 | 12/4/2017 |
0.11.2 | 174,001 | 10/24/2017 |
0.11.1 | 23,503 | 10/18/2017 |
0.11.0 | 189,206 | 7/19/2017 |
0.9.5 | 154,069 | 4/21/2017 |
0.9.4 | 33,598 | 3/1/2017 |