Ensurepass  QUESTION 51 Under which scenario would it be most appropriate to consider using faster (e.g 10 Gigabit) Ethernet as the network fabric for your Hadoop cluster?   A. When the typical workloads generates a large amount of intermediate data, on the order of the input data itself. B. When the typical workloads consists of processor-intensive tasks. C. When the typical workloads consumes a large amount of input data, relative to the entire capacity of HDFS. D. When Read more [...]
Ensurepass  QUESTION 21 Which command does Hadoop offer to discover missing or corrupt HDFS data?   A. The map-only checksum utility, B. Fsck C. Du D. Dskchk E. Hadoop does not provide any tools to discover missing or corrupt data; there is no need because three replicas are kept for each data block.   Answer: B Explanation: HDFS supports fsck command to check for various inconsistencies. It it is designed for reporting problems with various files, for e.g. missing Read more [...]
Ensurepass  QUESTION 31 You configure you cluster with HDFS High Availability (HA) using Quorum-Based storage. You do not implement HDFS Federation.   What is the maximum number of NameNodes daemon you should run on you cluster in order to avoid a "split-brain" scenario with your NameNodes?   A. Unlimited. HDFS High Availability (HA) is designed to overcome limitations on the number of NameNodes you can deploy. B. Two active NameNodes and one Standby NameNode C. One active Read more [...]
Ensurepass  QUESTION 11 You has a cluster running with the Fail Scheduler enabled. There are currently no jobs running on the cluster you submit a job A, so that only job A is running on the cluster. A while later, you submit job B.Now job A and Job B are running on the cluster al the same time. How will the Fair' Scheduler handle these two Jobs?   A. When job A gets submitted, it consumes all the task slot B. When job A gets submitted, it doesn't consume all the task slot C. Read more [...]
Ensurepass  Exam A   QUESTION 1 It is recommended that you run the HDFS balancer periodically. Why? (Choose 3)   A. To improve data locality for MapReduce tasks. B. To ensure that there is capacity in HDTS for additional data. C. To help HDFS deliver consistent performance under heavy loads. D. To ensure that all blocks in the cluster are 128MB in size. E. To ensure that there is consistent disk utilization across the DataNodes.   Answer: BCE Reference: http://hadoop.apache.org/docs/hdfs/r0.21.0/api/org/apache/hadoop/hdfs/server/balancer/Ba Read more [...]