Intermediate / 1 to 5 years experienced level questions
Intermediate / 1 to 5 years experienced level questions & answers
Ques 1. Explain the role of the offset.
There is a sequential ID number given to the messages in the partitions what we call, an offset. So, to identify each message in the partition uniquely, we use these offsets.
Ques 2. Explain the concept of Leader and Follower.
In every partition of Kafka, there is one server which acts as the Leader, and none or more servers plays the role as a Followers.
Ques 3. What is the process for starting a Kafka server?
- In order to start the ZooKeeper server: > bin/zookeeper-server-start.sh config/zookeeper.properties
- Next, to start the Kafka server: > bin/kafka-server-start.sh config/server.properties
Ques 4. Explain the role of the Kafka Producer API.
An API which permits an application to publish a stream of records to one or more Kafka topics is what we call Producer API.
Ques 5. What can you do with Kafka?
It can perform in several ways, such as:
- In order to transmit data between two systems, we can build a real-time stream of data pipelines with it.
- Also, we can build a real-time streaming platform with Kafka, that can actually react to the data.
Ques 6. What does ISR stand in Kafka environment?
ISR refers to In Sync Replicas. These are generally classified as a set of message replicas which are synced to be leaders.
Ques 7. What is the role of Consumer API?
An API which permits an application to subscribe to one or more topics and also to process the stream of records produced to them is what we call Consumer API.
Ques 8. Explain the role of Streams API in Kafka?
An API which permits an application to act as a stream processor, and also consuming an input stream from one or more topics and producing an output stream to one or more output topics, moreover, transforming the input streams to output streams effectively, is what we call Streams API.
Ques 9. What is the role of Connector API?
An API which permits to run as well as build the reusable producers or consumers which connect Kafka topics to existing applications or data systems is what we call the Connector API.
Ques 10. Why Should we use Apache Kafka Cluster?
- It is possible to track web activities just by storing/sending the events for real-time processes.
- Through this, we can Alert as well as report the operational metrics.
- Also, we can transform data into the standard format.
- Moreover, it allows continuous processing of streaming data to the topics.
Ques 11. What is Data Log in Kafka?
As we know, messages are retained for a considerable amount of time in Kafka. Moreover, there is flexibility for consumers that they can read as per their convenience. Although, there is a possible case that if Kafka is configured to keep messages for 24 hours and possibly that time consumer is down for time greater than 24 hours, then the consumer may lose those messages. However, still, we can read those messages from last known offset, but only at a condition that the downtime on part of the consumer is just 60 minutes. Moreover, on what consumers are reading from a topic Kafka doesn’t keep state.
Ques 12. Explain how to Tune Kafka for Optimal Performance.
- Tuning Kafka Producers
- Kafka Brokers Tuning
- Tuning Kafka Consumers
Ques 13. State Disadvantages of Apache Kafka.
- No Complete Set of Monitoring Tools.
- Issues with Message Tweaking.
- Not support wildcard topic selection.
- Lack of Pace.
Ques 14. Enlist all Apache Kafka Operations.
Apache Kafka Operations are:
- Addition and Deletion of Kafka Topics
- How to modify the Kafka Topics
- Distinguished Turnoff
- Mirroring Data between Kafka Clusters
- Finding the position of the Consumer
- Expanding Your Kafka Cluster
- Migration of Data Automatically
- Retiring Servers
- Datacenters
Ques 15. Explain Apache Kafka Use Cases?
- Kafka Metrics: It is possible to use Kafka for operational monitoring data. Also, to produce centralized feeds of operational data, it involves aggregating statistics from distributed applications.
- Kafka Log Aggregation: Moreover, to gather logs from multiple services across an organization.
- Stream Processing: While stream processing, Kafka’s strong durability is very useful.
Ques 16. What role does ZooKeeper play in a cluster of Kafka?
Apache ZooKeeper acts as a distributed, open-source configuration and synchronization service, along with being a naming registry for distributed applications. It keeps track of the status of the Kafka cluster nodes, as well as of Kafka topics, partitions, etc.
Since the data is divided across collections of nodes within ZooKeeper, it exhibits high availability and consistency. When a node fails, ZooKeeper performs an instant failover migration.
ZooKeeper is used in Kafka for managing service discovery for Kafka brokers, which form the cluster. ZooKeeper communicates with Kafka when a new broker joins, when a broker dies, when a topic gets removed, or when a topic is added so that each node in the cluster knows about these changes. Thus, it provides an in-sync view of the Kafka cluster configuration.
Ques 17. Elaborate the architecture of Kafka.
In Kafka, a cluster contains multiple brokers since it is a distributed system. Topic in the system will get divided into multiple partitions, and each broker stores one or more of those partitions so that multiple producers and consumers can publish and retrieve messages at the same time.
Most helpful rated by users:
- Explain the concept of Leader and Follower.
- Enlist the several components in Kafka.
- What are main APIs of Kafka?