Question
a.
Partitioner
b.
OutputSplit
c.
Reporter
d.
All of the mentioned
Posted under Hadoop
Engage with the Community - Add Your Comment
Confused About the Answer? Ask for Details Here.
Know the Explanation? Add it Here.
Q. Users can control which keys (and hence records) go to which Reducer by implementing a custom?
Similar Questions
Discover Related MCQs
Q. Applications can use the ____________ to report progress and set application-level status messages.
View solution
Q. The right level of parallelism for maps seems to be around _________ maps per-node.
View solution
Q. The number of reduces for the job is set by the user via _________
View solution
Q. The framework groups Reducer inputs by key in _________ stage.
View solution
Q. The output of the reduce task is typically written to the FileSystem via _____________
View solution
Q. Which of the following is the default Partitioner for Mapreduce?
View solution
Q. Which of the following partitions the key space?
View solution
Q. ____________ is a generalization of the facility provided by the MapReduce framework to collect data output by the Mapper or the Reducer.
View solution
Q. __________ is the primary interface for a user to describe a MapReduce job to the Hadoop framework for execution.
View solution
Q. The ___________ executes the Mapper/ Reducer task as a child process in a separate jvm.
View solution
Q. Maximum virtual memory of the launched child-task is specified using _________
View solution
Q. Which of the following parameter is the threshold for the accounting and serialization buffers?
View solution
Q. ______________ is percentage of memory relative to the maximum heap size in which map outputs may be retained during the reduce.
View solution
Q. ____________ specifies the number of segments on disk to be merged at the same time.
View solution
Q. Map output larger than ___________ percent of the memory allocated to copying map outputs.
View solution
Q. Jobs can enable task JVMs to be reused by specifying the job configuration _________
View solution
Q. During the execution of a streaming job, the names of the _______ parameters are transformed.
View solution
Q. The standard output (stdout) and error (stderr) streams of the task are read by the TaskTracker and logged to _________
View solution
Q. ____________ is the primary interface by which user-job interacts with the JobTracker.
View solution
Q. The _____________ can also be used to distribute both jars and native libraries for use in the map and/or reduce tasks.
View solution
Suggested Topics
Are you eager to expand your knowledge beyond Hadoop? We've curated a selection of related categories that you might find intriguing.
Click on the categories below to discover a wealth of MCQs and enrich your understanding of Computer Science. Happy exploring!