Workshop: Apache Kafka and KSQL in Action : Let’s Build a Streaming Data Pipeline! Part 1/2

Have you ever thought that you needed to be a programmer to do stream processing and build streaming data pipelines? Think again! Apache Kafka is a distributed, scalable, and fault-tolerant streaming platform, providing low-latency pub-sub messaging coupled with native storage and stream processing capabilities. Integrating Kafka with RDBMS, NoSQL, and object stores is simple with Kafka Connect, which is part of Apache Kafka. KSQL is the open-source SQL streaming engine for Apache Kafka, and makes it possible to build stream processing applications at scale, written using a familiar SQL interface.

In this workshop you will hear the architectural reasoning for Apache Kafka and the benefits of real-time integration, and then build a streaming data pipeline using nothing but your bare hands, Kafka Connect, and KSQL.

Gasp as we filter events in real time! Be amazed at how we can enrich streams of data with data from RDBMS! Be astonished at the power of streaming aggregates for anomaly detection!

Prerequisite
Participants must have their own laptop and complete the prerequisite steps detailed here PRIOR TO THE WORKSHOP:https://github.com/confluentinc/demo-scene/blob/master/ksql-workshop/pre-requisites.adoc.

Computer Setup
For reasons of compatibility, we recommend the use of Mac or Linux for the workshop. 8GB memory is required.Knowledge of SQL is assumed. Participants will benefit from existing knowledge of the concepts of Kafka (topics, streaming, etc).