Pass the Cloudera CCAH CCA-500 Questions and answers with CertsForce

Viewing page 2 out of 2 pages
Viewing questions 11-20 out of questions
Questions # 11:

You are running a Hadoop cluster with a NameNode on host mynamenode, a secondary NameNode on host mysecondarynamenode and several DataNodes.

Which best describes how you determine when the last checkpoint happened?

Options:

A.

Execute hdfs namenode –report on the command line and look at the Last Checkpoint information


B.

Execute hdfs dfsadmin –saveNamespace on the command line which returns to you the last checkpoint value in fstime file


C.

Connect to the web UI of the Secondary NameNode (http://mysecondary:50090/) and look at the “Last Checkpoint” information


D.

Connect to the web UI of the NameNode (http://mynamenode:50070) and look at the “Last Checkpoint” information


Expert Solution
Questions # 12:

Your cluster implements HDFS High Availability (HA). Your two NameNodes are named nn01 and nn02. What occurs when you execute the command: hdfs haadmin –failover nn01 nn02?

Options:

A.

nn02 is fenced, and nn01 becomes the active NameNode


B.

nn01 is fenced, and nn02 becomes the active NameNode


C.

nn01 becomes the standby NameNode and nn02 becomes the active NameNode


D.

nn02 becomes the standby NameNode and nn01 becomes the active NameNode


Expert Solution
Questions # 13:

You use the hadoop fs –put command to add a file “sales.txt” to HDFS. This file is small enough that it fits into a single block, which is replicated to three nodes in your cluster (with a replication factor of 3). One of the nodes holding this file (a single block) fails. How will the cluster handle the replication of file in this situation?

Options:

A.

The file will remain under-replicated until the administrator brings that node back online


B.

The cluster will re-replicate the file the next time the system administrator reboots the NameNode daemon (as long as the file’s replication factor doesn’t fall below)


C.

This will be immediately re-replicated and all other HDFS operations on the cluster will halt until the cluster’s replication values are resorted


D.

The file will be re-replicated automatically after the NameNode determines it is under-replicated based on the block reports it receives from the NameNodes


Expert Solution
Questions # 14:

Which three basic configuration parameters must you set to migrate your cluster from MapReduce 1 (MRv1) to MapReduce V2 (MRv2)? (Choose three)

Options:

A.

Configure the NodeManager to enable MapReduce services on YARN by setting the following property in yarn-site.xml:

yarn.nodemanager.hostname

your_nodeManager_shuffle


B.

Configure the NodeManager hostname and enable node services on YARN by setting the following property in yarn-site.xml:

yarn.nodemanager.hostname

your_nodeManager_hostname


C.

Configure a default scheduler to run on YARN by setting the following property in mapred-site.xml:

mapreduce.jobtracker.taskScheduler

org.apache.hadoop.mapred.JobQueueTaskScheduler


D.

Configure the number of map tasks per jon YARN by setting the following property in mapred:

mapreduce.job.maps

2


E.

Configure the ResourceManager hostname and enable node services on YARN by setting the following property in yarn-site.xml:

yarn.resourcemanager.hostname

your_resourceManager_hostname


F.

Configure MapReduce as a Framework running on YARN by setting the following property in mapred-site.xml:

mapreduce.framework.name

yarn


Expert Solution
Questions # 15:

You have a cluster running with a FIFO scheduler enabled. You submit a large job A to the cluster, which you expect to run for one hour. Then, you submit job B to the cluster, which you expect to run a couple of minutes only.

You submit both jobs with the same priority.

Which two best describes how FIFO Scheduler arbitrates the cluster resources for job and its tasks? (Choose two)

Options:

A.

Because there is a more than a single job on the cluster, the FIFO Scheduler will enforce a limit on the percentage of resources allocated to a particular job at any given time


B.

Tasks are scheduled on the order of their job submission


C.

The order of execution of job may vary


D.

Given job A and submitted in that order, all tasks from job A are guaranteed to finish before all tasks from job B


E.

The FIFO Scheduler will give, on average, and equal share of the cluster resources over the job lifecycle


F.

The FIFO Scheduler will pass an exception back to the client when Job B is submitted, since all slots on the cluster are use


Expert Solution
Questions # 16:

You want to node to only swap Hadoop daemon data from RAM to disk when absolutely necessary. What should you do?

Options:

A.

Delete the /dev/vmswap file on the node


B.

Delete the /etc/swap file on the node


C.

Set the ram.swap parameter to 0 in core-site.xml


D.

Set vm.swapfile file on the node


E.

Delete the /swapfile file on the node


Expert Solution
Questions # 17:

You have A 20 node Hadoop cluster, with 18 slave nodes and 2 master nodes running HDFS High Availability (HA). You want to minimize the chance of data loss in your cluster. What should you do?

Options:

A.

Add another master node to increase the number of nodes running the JournalNode which increases the number of machines available to HA to create a quorum


B.

Set an HDFS replication factor that provides data redundancy, protecting against node failure


C.

Run a Secondary NameNode on a different master from the NameNode in order to provide automatic recovery from a NameNode failure.


D.

Run the ResourceManager on a different master from the NameNode in order to load-share HDFS metadata processing


E.

Configure the cluster’s disk drives with an appropriate fault tolerant RAID level


Expert Solution
Questions # 18:

What two processes must you do if you are running a Hadoop cluster with a single NameNode and six DataNodes, and you want to change a configuration parameter so that it affects all six DataNodes. (Choose two)

Options:

A.

You must modify the configuration files on the NameNode only. DataNodes read their configuration from the master nodes


B.

You must modify the configuration files on each of the DataNodes machines


C.

You don’t need to restart any daemon, as they will pick up changes automatically


D.

You must restart the NameNode daemon to apply the changes to the cluster


E.

You must restart all six DatNode daemon to apply the changes to the cluster


Expert Solution
Viewing page 2 out of 2 pages
Viewing questions 11-20 out of questions