Stream processing with Kafka
Tim Berglund leads a basic architectural introduction to Kafka and walks you through using Kafka Streams and KSQL to process streaming data.
Talk Title | Stream processing with Kafka |
Speakers | Tim Berglund (Confluent) |
Conference | Strata Data Conference |
Conf Tag | Big Data Expo |
Location | San Jose, California |
Date | March 6-8, 2018 |
URL | Talk Page |
Slides | Talk Slides |
Video | |
The toolset for building scalable data systems is maturing, having adapted well to our decades-old paradigm of update-in-place databases. We ingest events, we store them in high-volume OLTP databases, and we have new OLAP systems to analyze them at scale—even if the size of our operation requires us to grow to dozens or hundreds of servers in the distributed system. But something feels a little dated about the store-and-analyze paradigm, as if we are missing a new architectural insight that might more efficiently distribute the work of storing and computing the events that happen to our software. That new paradigm is stream processing. Tim Berglund leads a basic architectural introduction to Apache Kafka and walks you through using Kafka Streams and KSQL to process streaming data. You’ll learn the basics of Kafka as a messaging system, including the core concepts of topic, producer, consumer, and broker. Tim also explains how topics are partitioned among brokers and highlights the simple Java APIs for getting data in and out. But more importantly, you’ll discover how to extend this scalable messaging system into a streaming data processing system—one that offers significant advantages in scalability and deployment agility while locating computation in your data pipeline in precisely the places it belongs: in your microservices and applications, not in costly, high-density systems.