Hadoop Questions and Answers Part-22

1. Which of the following method is used to get user-specified job name?
a) getJobName()
b) getJobState()
c) getPriority()
d) all of the mentioned

Answer: a
Explanation: getPriority() is used to get scheduling info of the job.

2. Which of the following is used to list out the storage policies?
a) hdfs storagepolicies
b) hdfs storage
c) hd storagepolicies
d) all of the mentioned

Answer: a
Explanation: Arguments are none for the hdfs storagepolicies command.

3. Which of the following statement can be used to get the storage policy of a file or a directory?
a) hdfs dfsadmin -getStoragePolicy path
b) hdfs dfsadmin -setStoragePolicy path policyName
c) hdfs dfsadmin -listStoragePolicy path policyName
d) all of the mentioned

Answer: a
Explanation: refers to the path referring to either a directory or a file.

4. __________ get events indicating completion (success/failure) of component tasks.
a) getJobName()
b) getJobState()
c) getPriority()
d) getTaskCompletionEvents(int startFrom)

Answer: d
Explanation: getPriority() provides scheduling info of the job.

5. _________ gets the diagnostic messages for a given task attempt.
a) getTaskOutputFilter(Configuration conf)
b) getTaskReports(TaskType type)
c) getTrackingURL()
d) all of the mentioned

Answer: a
Explanation: getTaskDiagnostics(TaskAttemptID taskid) gets the diagnostic messages for a given task attempt.

6. reduceProgress() gets the progress of the job’s reduce-tasks, as a float between _________
a) 0.0-1.0
b) 1.0-2.0
c) 2.0-3.0
d) None of the mentioned

Answer: a
Explanation: mapProgress() is used to get the progress of the job’s map-tasks, as a float between 0.0 and 1.0.

7. The Job makes a copy of the _____________ so that any necessary internal modifications do not reflect on the incoming parameter.
a) Component
b) Configuration
c) Collector
d) None of the mentioned

Answer: b
Explanation: A Cluster will be created from the conf parameter only when it’s needed.

8. ________ is the architectural center of Hadoop that allows multiple data processing engines.
a) YARN
b) Hive
c) Incubator
d) Chuckwa

Answer: a
Explanation: YARN is the prerequisite for Enterprise Hadoop, providing resource management and a central platform to deliver consistent operations, security, and data governance tools across Hadoop clusters.

9. Point out the correct statement.
a) YARN also extends the power of Hadoop to incumbent and new technologies found within the data center
b) YARN is the central point of investment for Hortonworks within the Apache community
c) YARN enhances a Hadoop compute cluster in many ways
d) All of the mentioned

Answer: d
Explanation: YARN provides ISVs and developers a consistent framework for writing data access applications that run IN Hadoop.

10. YARN’s dynamic allocation of cluster resources improves utilization over more static _______ rules used in early versions of Hadoop.
a) Hive
b) MapReduce
c) Imphala
d) All of the mentioned

Answer: b
Explanation: Multi-tenant data processing improves an enterprise’s return on its Hadoop investments.