Get started with RBAC at scale, Oracle CDC Source Connector, and more within our Q2 Launch for Confluent Cloud | Register for demo

How Apache Kafka® Works

Pick up best practices for developing applications that use Apache Kafka, beginning with a high level code overview for a basic producer and consumer. From there we’ll cover strategies for building powerful stream processing applications, including high availability through replication, data retention policies, producer design and producer guarantees.

We’ll delve into the details of delivery guarantees, including exactly-once semantics, partition strategies and consumer group rebalances. The talk will finish with a discussion of compacted topics, troubleshooting strategies and a security overview.

This session is part 3 of 4 in our Fundamentals for Apache Kafka series.

<<Back