Concepts of Apache Kafka

Saurav Kumar
11 min readSep 25, 2023

Apache Kafka is a popular open-source stream processing platform that is widely used for building real-time data pipelines and event-driven applications. It was originally developed by LinkedIn and later open-sourced as an Apache project. Kafka is designed to handle high-throughput, fault-tolerant, scalable data streaming and it relies on a set of internal components and mechanisms to achieve these goals. Here are some key concepts associated with Apache Kafka and an overview of how Kafka works internally:

  1. Topics:
  • In Kafka, data is organized into topics. A topic is a logical channel or category to which messages are published by producers and from which messages are consumed by consumers.
  • Topics allow you to categorize and organize the data streams based on different data sources, events, or use cases.

2. Producer:

  • Producers are applications or systems that push data into Kafka topics.
  • They are responsible for creating and publishing messages to Kafka topics.
  • Producers can be configured to send messages to one or more Kafka topics.

3. Consumer:

  • Consumers are applications or systems that subscribe to Kafka topics and process messages.
  • Consumers can read messages from one or…

--

--

Saurav Kumar
Saurav Kumar

Written by Saurav Kumar

Experienced Software Engineer adept in Java, Spring Boot, Microservices, Kafka & Azure.

No responses yet