An open-source system for data streaming in which a broker queues and persists data records (also called events or messages) between sources (publishers) and targets (subscribers). Sequences of these records are sorted into streams called topics.
Added Perspectives
In a nutshell, Kafka “producers” publish data records to brokers that persist those records to file systems on disk until they are read – either real-time or later on – by “consumers.” Records are divided into topics (a.k.a. streams) to which consumers can subscribe for selected use. Topics also can be partitioned to improve throughput via parallel consumption and enable redundancy via partition replicas across multiple clustered brokers. Here is a sample architecture for one topic. While this shows multiple examples of producers, one topic often maps to a single producer.
- Kevin Petrie in Assessing the Kafka Data Streaming Opportunity
September 17, 2018 (Blog)
Relevant Content
Nov 06, 2018 - With Kafka, enterprises can address new advanced analytics use cases. But does it displace data architecture pillars, such as the database?
Sep 06, 2019 - Learn how to achieve the DataOps objectives of improved efficiency and data quality by migrating to a streaming architecture based on Apache Kafka.
Related Terms
Unleash The Power Of Your Data
Providing modern, comprehensive data solutions so you can turn data into your most powerful asset and stay ahead of the competition.
Learn how we can help your organization create actionable data strategies and highly tailored solutions.
© Datalere, LLC. All rights reserved