JOBSEEKERS
Login
Sign Up
Jobseeker
Employer
Staffing Firm
Direct Client
Hadoop interview questions part 47
Hadoop interview questions part 47
Back
Take as many assements as you can to improve your validate your skill rating
Total Questions: 20
1. Sqoop can also import the data into Hive by generating and executing a ____________ statement to define the data’s layout in Hive.
A. SET TABLE
B. CREATE TABLE
C. INSERT TABLE
D. All of the mentioned
Show Correct Answer
Correct Answer is :
CREATE TABLE
2. The __________ tool imports a set of tables from an RDBMS to HDFS.
A. export-all-tables
B. import-all-tables
C. import-tables
D. none of the mentioned
Show Correct Answer
Correct Answer is :
import-tables
3. Which of the following argument is not supported by import-all-tables tool ?
A. –class-name
B. –package-name
C. –database-name
D. –table-name
Show Correct Answer
Correct Answer is :
–class-name
4. ____________ is a distributed real-time computation system for processing large volumes of high-velocity data.
A. Kafka
B. Storm
C. Lucene
D. BigTop
Show Correct Answer
Correct Answer is :
Storm
5. Point out the correct statement :
A. A Storm topology consumes streams of data and processes those streams in arbitrarily complex ways
B. Apache Storm is a free and open source distributed real-time computation system
C. Storm integrates with the queueing and database technologies you already use
D. All of the mentioned
Show Correct Answer
Correct Answer is :
All of the mentioned
6. Storm integrates with __________ via Apache Slider
A. Scheduler
B. YARN
C. Compaction
D. All of the mentioned
Show Correct Answer
Correct Answer is :
Compaction
7. For Apache __________ users, Storm utilizes the same ODBC interface.
A. cTakes
B. Hive
C. Pig
D. Oozie
Show Correct Answer
Correct Answer is :
Hive
8. Point out the wrong statement :
A. Storm is difficult and can be used with only Java
B. Storm is fast: a benchmark clocked it at over a million tuples processed per second per node
C. Storm is scalable, fault-tolerant, guarantees your data will be processed
D. All of the mentioned
Show Correct Answer
Correct Answer is :
Storm is difficult and can be used with only Java
9. Storm is benchmarked as processing one million _______ byte messages per second per node
A. 10
B. 50
C. 100
D. 200
Show Correct Answer
Correct Answer is :
100
10. Apache Storm added open source, stream data processing to _________ Data Platform
A. Cloudera
B. Hortonworks
C. Local Cloudera
D. MapR
Show Correct Answer
Correct Answer is :
Hortonworks
11. How many types of nodes are present in Storm cluster ?
A. 1
B. 2
C. 3
D. 4
Show Correct Answer
Correct Answer is :
3
12. __________ node distributes code across the cluster.
A. Zookeeper
B. Nimbus
C. Supervisor
D. None of the mentioned
Show Correct Answer
Correct Answer is :
Nimbus
13. ____________ communicates with Nimbus through Zookeeper, starts and stops workers according to signals from Nimbus
A. Zookeeper
B. Nimbus
C. Supervisor
D. None of the mentioned
Show Correct Answer
Correct Answer is :
Supervisor
14. Which of the following node is responsible for executing a Task assigned to it by the JobTracker ?
A. MapReduce
B. Mapper
C. TaskTracker
D. JobTracker
Show Correct Answer
Correct Answer is :
TaskTracker
15. Point out the correct statement :
A. MapReduce tries to place the data and the compute as close as possible
B. Map Task in MapReduce is performed using the Mapper() function
C. Reduce Task in MapReduce is performed using the Map() function
D. All of the mentioned
Show Correct Answer
Correct Answer is :
MapReduce tries to place the data and the compute as close as possible
16. ___________ part of the MapReduce is responsible for processing one or more chunks of data and producing the output results.
A. Maptask
B. Mapper
C. Task execution
D. All of the mentioned
Show Correct Answer
Correct Answer is :
Maptask
17. _________ function is responsible for consolidating the results produced by each of the Map() functions/tasks.
A. Reduce
B. Map
C. Reducer
D. All of the mentioned
Show Correct Answer
Correct Answer is :
Reduce
18. Point out the wrong statement :
A. A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner
B. The MapReduce framework operates exclusively on pairs
C. Applications typically implement the Mapper and Reducer interfaces to provide the map and reduce methods
D. None of the mentioned
Show Correct Answer
Correct Answer is :
None of the mentioned
19. Although the Hadoop framework is implemented in Java , MapReduce applications need not be written in :
A. Java
B. C
C. C#
D. None of the mentioned
Show Correct Answer
Correct Answer is :
Java
20. ________ is a utility which allows users to create and run jobs with any executable as the mapper and/or the reducer.
A. Hadoop Strdata
B. Hadoop Streaming
C. Hadoop Stream
D. None of the mentioned
Show Correct Answer
Correct Answer is :
Hadoop Streaming
Similar Interview Questions
Search for latest jobs
Find Jobs