Dive into Avro Everything a data engineer needs to know - Ryan Skraba, Ismaël Mejía
A presentation from ApacheCon 2022
https://apachecon.com/acna2022/slides/02_SkrabaMejia_Dive_into_Avro.pdf
Apache Avro is the de-facto standard for serializing structured data: Big data engines, streaming platforms, and data lakes use it to optimise storage or transmission of data.
Avro is stable and mature, and usually "just works", but there's a lot going on behind the scenes. This talk covers topics, gotchas and best practices for using Avro in your systems: types and schemas, logical types, binary and JSON serialization, code generation, and its strong ecosystem of tools.
Finally we'll dive a bit deeper into schema evolution, an important topic especially in event-driven architectures.