White Paper

Kafka Serialization and Deserialization (SerDes) Examples

Download the White Paper

To put real-time data to work, event streaming applications rely on stream processing, Kafka Connect allows developers to capture events from end systems.. By leveraging Kafka Streams API as well, developers can build pipelines that both ingest and transform streaming data for application consumption, designed with data stream format and serialization in mind.

Download the white paper to explore five examples of different data formats, SerDes combinations connector configurations and Kafka Streams code for building event streaming pipelines:

  1. Confluent CLI Producer with String
  2. JDBC source connector with JSON
  3. JDBC source connector with SpecificAvro
  4. JDBC source connector with GenericAvro
  5. Java producer with SpecificAvro

Yeva is an integration architect at Confluent designing solutions and building demos for developers and operators of Apache Kafka. She has many years of experience validating and optimizing end-to-end solutions for distributed software systems and networks.

その他のリソース

cc demo

Confluent Cloud Demo

Join us for a live demo of Confluent Cloud, the industry’s only fully managed, cloud-native event streaming platform powered by Apache Kafka
kafka microservices

Kafka Microservices

In this online talk series, learn key concepts, use cases and best practices to harness the power of real-time streams for microservices architectures
Image-Event-Driven Microservices-01

e-book: Microservices Customer Stories

See how five organizations across a wide range of industries leveraged Confluent to build a new class of event-driven microservices