This set of Hadoop Multiple Choice Questions & Answers (MCQs) focuses on “Kafka with Hadoop – 1”.
1. Kafka is comparable to traditional messaging systems such as _____________
a) Impala
b) ActiveMQ
c) BigTop
d) Zookeeper
View Answer
Explanation: Kafka works well as a replacement for a more traditional message broker.
2. Point out the correct statement.
a) The original use case for Kafka was to be able to rebuild a user activity tracking pipeline as a set of real-time publish-subscribe feeds
b) Activity tracking is often very high volume as many activity messages are generated for each user page view
c) Kafka is often used for operational monitoring data
d) All of the mentioned
View Answer
Explanation: This involves aggregating statistics from distributed applications to produce centralized feeds of operational data.
3. Many people use Kafka as a replacement for a ___________ solution.
a) log aggregation
b) compaction
c) collection
d) all of the mentioned
View Answer
Explanation: Log aggregation typically collects physical log files off servers and puts them in a central place.
4. _______________ is a style of application design where state changes are logged as a time-ordered sequence of records.
a) Event sourcing
b) Commit Log
c) Stream Processing
d) None of the mentioned
View Answer
Explanation: Kafka’s support for very large stored log data makes it an excellent backend for an application built in this style.
5. Point out the wrong statement.
a) Kafka can serve as a kind of external commit-log for a distributed system
b) The log helps replicate data between nodes and acts as a re-syncing mechanism for failed nodes to restore their data
c) Kafka comes with a command-line client that will take input from a file or from standard input and send it out as messages to the Kafka cluster
d) All of the mentioned
View Answer
Explanation: By default, each line will be sent as a separate message.
6. Kafka uses __________ so you need to first start a ZooKeeper server if you don’t already have one.
a) Impala
b) ActiveMQ
c) BigTop
d) Zookeeper
View Answer
Explanation: You can use the convenience script packaged with Kafka to get a quick-and-dirty single-node ZooKeeper instance.
7. __________ is the node responsible for all reads and writes for the given partition.
a) replicas
b) leader
c) follower
d) isr
View Answer
Explanation: Each node will be the leader for a randomly selected portion of the partitions.
8. __________ is the subset of the replicas list that is currently alive and caught up to the leader.
a) replicas
b) leader
c) follower
d) isr
View Answer
Explanation: “isr” is the set of “in-sync” replicas.
9. Kafka uses key-value pairs in the ____________ file format for configuration.
a) RFC
b) Avro
c) Property
d) None of the mentioned
View Answer
Explanation: These key values can be supplied either from a file or programmatically.
10. __________ is the amount of time to keep a log segment before it is deleted.
a) log.cleaner.enable
b) log.retention
c) log.index.enable
d) log.flush.interval.message
View Answer
Explanation: log.cleaner.enable is configuration must be set to true for log compaction to run.
Sanfoundry Global Education & Learning Series – Hadoop.
Here’s the list of Best Books in Hadoop.
- Get Free Certificate of Merit in Hadoop
- Participate in Hadoop Certification Contest
- Become a Top Ranker in Hadoop
- Take Hadoop Tests
- Chapterwise Practice Tests: Chapter 1, 2, 3, 4, 5, 6, 7, 8, 9, 10
- Chapterwise Mock Tests: Chapter 1, 2, 3, 4, 5, 6, 7, 8, 9, 10
- Apply for Hadoop Internship
- Apply for Information Technology Internship
- Buy Programming Books
- Buy Hadoop Books
- Practice Programming MCQs