Evolve Your Schemas in a Better Way! A Deep Dive into Avro Schema Compatibility and Schema Registry. The only constant in life is change! The same applies to your Kafka events flowing through your streaming applications.
The Confluent Schema Registry allows us to control how schemas can evolve over time without breaking the compatibility of our streaming applications. But when you start with Kafka and (Avro) schemas, this can be pretty overwhelming.
Join Kosta and Tim as we dive into the tricky world of backward and forward compatibility in schema design. During this deep dive talk, we are going to answer questions like:
* What compatibility level to pick?
* What changes can I make when evolving my schemas?
* What options do I have when I need to introduce a breaking change?
* Should we automatically register schemas from our applications? Or do we need a separate step in our deployment process to promote schemas to higher-level environments?
* What to promote first? My producer, consumer or schema?
* How do you generate Java classes from your Avro schemas using Maven or Gradle, and how to integrate this into your project(s)?
* How do you build an automated test suite (unit tests) to gain more confidence and verify you are not breaking compatibility? Even before deploying a new version of your schema or application.
With live demos, we'll show you how to make schema changes work seamlessly. Emphasizing the crucial decisions, using real-life examples, pitfalls and best practices when promoting schemas on the consumer and producer sides.
Explore the ins and outs of Apache Avro and the Schema Registry with us at the Kafka Summit! Start evolving your schemas in a better way today, and join this talk!