JOBSEEKERS
Login
Sign Up
Jobseeker
Employer
Staffing Firm
Direct Client
Hadoop interview questions part 30
Hadoop interview questions part 30
Back
Take as many assements as you can to improve your validate your skill rating
Total Questions: 20
1. ______________ is method to copy byte from input stream to any other stream in Hadoop.
A. IOUtils
B. Utils
C. IUtils
D. All of the mentioned
Show Correct Answer
Correct Answer is :
IOUtils
2. _____________ is used to read data from bytes buffers .
A. write()
B. read()
C. readwrite()
D. all of the mentioned
Show Correct Answer
Correct Answer is :
write()
3. Point out the wrong statement :
A. The framework calls reduce method for each pair in the grouped inputs
B. The output of the Reducer is re-sorted
C. reduce method reduces values for a given key
D. None of the mentioned
Show Correct Answer
Correct Answer is :
The output of the Reducer is re-sorted
4. Interface ____________ reduces a set of intermediate values which share a key to a smaller set of values.
A. Mapper
B. Reducer
C. Writable
D. Readable
Show Correct Answer
Correct Answer is :
Reducer
5. Reducer is input the grouped output of a :
A. Mapper
B. Reducer
C. Writable
D. Readable
Show Correct Answer
Correct Answer is :
Mapper
6. The output of the reduce task is typically written to the FileSystem via :
A. OutputCollector
B. InputCollector
C. OutputCollect
D. All of the mentioned
Show Correct Answer
Correct Answer is :
OutputCollector
7. Applications can use the _________ provided to report progress or just indicate that they are alive.
A. Collector
B. Reporter
C. Dashboard
D. None of the mentioned
Show Correct Answer
Correct Answer is :
Reporter
8. Which of the following parameter is to collect keys and combined values ?
A. key
B. values
C. reporter
D. output
Show Correct Answer
Correct Answer is :
output
9. Kafka is comparable to traditional messaging systems such as :
A. Impala
B. ActiveMQ
C. BigTop
D. Zookeeper
Show Correct Answer
Correct Answer is :
ActiveMQ
10. Point out the correct statement :
A. The original use case for Kafka was to be able to rebuild a user activity tracking pipeline as a set of real-time publish-subscribe feeds
B. Activity tracking is often very high volume as many activity messages are generated for each user page view.
C. Kafka is often used for operational monitoring data
D. All of the mentioned
Show Correct Answer
Correct Answer is :
All of the mentioned
11. Many people use Kafka as a replacement for a ___________ solution.
A. log aggregation
B. compaction
C. collection
D. all of the mentioned
Show Correct Answer
Correct Answer is :
log aggregation
12. _______________ is a style of application design where state changes are logged as a time-ordered sequence of records.
A. Event sourcing
B. Commit Log
C. Stream Processing
D. None of the mentioned
Show Correct Answer
Correct Answer is :
Event sourcing
13. Point out the wrong statement :
A. Kafka can serve as a kind of external commit-log for a distributed system
B. The log helps replicate data between nodes and acts as a re-syncing mechanism for failed nodes to restore their data
C. Kafka comes with a command line client that will take input from a file or from standard input and send it out as messages to the Kafka cluster
D. All of the mentioned
Show Correct Answer
Correct Answer is :
All of the mentioned
14. Kafka uses __________ so you need to first start a ZooKeeper server if you don’t already have one.
A. Impala
B. ActiveMQ
C. BigTop
D. Zookeeper
Show Correct Answer
Correct Answer is :
Zookeeper
15. __________ is the node responsible for all reads and writes for the given partition.
A. replicas
B. leader
C. follower
D. isr
Show Correct Answer
Correct Answer is :
leader
16. __________ is the subset of the replicas list that is currently alive and caught-up to the leader.
A. replicas
B. leader
C. follower
D. isr
Show Correct Answer
Correct Answer is :
isr
17. Kafka uses key-value pairs in the ____________ file format for configuration.
A. RFC
B. Avro
C. Property
D. None of the mentioned
Show Correct Answer
Correct Answer is :
Property
18. __________ is the amount of time to keep a log segment before it is deleted.
A. log.cleaner.enable
B. log.retention
C. log.index.enable
D. log.flush.interval.message
Show Correct Answer
Correct Answer is :
log.retention
19. __________ provides the functionality of a messaging system.
A. Oozie
B. Kafka
C. Lucene
D. BigTop
Show Correct Answer
Correct Answer is :
Kafka
20. Point out the correct statement :
A. With kafka, more users, whether using SQL queries or BI applications, can interact with more data
B. A topic is a category or feed name to which messages are published
C. For each topic, the Kafka cluster maintains a partitioned log
D. None of the mentioned
Show Correct Answer
Correct Answer is :
A topic is a category or feed name to which messages are published
Similar Interview Questions
Search for latest jobs
Find Jobs