A modern business operates 24/7 and generates data continuously. Shouldn’t we process it continuously too?
A rich ecosystem of real-time data-processing frameworks, tools and systems has been forming around Apache Kafka that allows data to be processed continuously as it occurs. Jay Kreps will introduce Kafka and explain why it has become the de facto standard for streaming data. He will draw on practical experience building stream-processing applications to discuss the difference between architectures and the challenges each presents. Jay will then outline the Kafka Streams API, which offers new stream processing functionality in Kafka, and explain how it helps tame some of the complexity in real-time architectures.
Visit www.confluent.io for more information