Project Metamorphosis: Unveiling the next-gen event streaming platformLearn More

Kafka Serialization and Deserialization (SerDes) Examples

Download the White Paper

Kafka Serialization and Deserialization (SerDes) Examples

Developers writing event streaming applications can use Kafka Connect to capture events from end systems and then use the Kafka Streams API to transform that data. In building these pipelines, they need to consider data stream format and serialization. For example, some Kafka Streams methods require record keys to be non-null, so either the connector or the application may need to add keys to the original event stream. Or, another consideration is how the record keys or record values are serialized—you must use the appropriate serializer/deserializer (SerDes) to convert data in each direction.

Download the white paper to dive into full Kafka examples, with connector configurations and Kafka Streams code, that demonstrate different data formats and SerDes combinations for building event streaming pipelines:

Example 1: Confluent CLI Producer with String
Example 2: JDBC source connector with JSON
Example 3: JDBC source connector with SpecificAvro
Example 4: JDBC source connector with GenericAvro
Example 5: Java producer with SpecificAvro

Author

Yeva Byzek

Yeva Byzek, Integration Architect, Confluent

Yeva is an integration architect at Confluent designing solutions and building demos for developers and operators of Apache Kafka. She has many years of experience validating and optimizing end-to-end solutions for distributed software systems and networks.

Sign Up Now

Start your 3-month trial. Get up to $200 off on each of your first 3 Confluent Cloud monthly bills

New signups only.

By clicking “sign up” above you understand we will process your personal information in accordance with our Privacy Policy.

By clicking "sign up" above you agree to the Terms of Service and to receive occasional marketing emails from Confluent. You also understand that we will process your personal information in accordance with our Privacy Policy.

Free Forever on a Single Kafka Broker
i

The software will allow unlimited-time usage of commercial features on a single Kafka broker. Upon adding a second broker, a 30-day timer will automatically start on commercial features, which cannot be reset by moving back to one broker.

Select Deployment Type
Manual Deployment
  • tar
  • zip
  • deb
  • rpm
  • docker
or
Auto Deployment
  • kubernetes
  • ansible

By clicking "download free" above you understand we will process your personal information in accordance with our Privacy Policy.

By clicking "download free" above, you agree to the Confluent License Agreement and to receive occasional marketing emails from Confluent. You also agree that your personal data will be processed in accordance with our Privacy Policy.

This website uses cookies to enhance user experience and to analyze performance and traffic on our website. We also share information about your use of our site with our social media, advertising, and analytics partners.