Exam Details

  • Exam Code
    :CCA-500
  • Exam Name
    :Cloudera Certified Administrator for Apache Hadoop (CCAH)
  • Certification
    :CCAH
  • Vendor
    :Cloudera
  • Total Questions
    :60 Q&As
  • Last Updated
    :May 14, 2024

Cloudera CCAH CCA-500 Questions & Answers

  • Question 11:

    You have a Hadoop cluster HDFS, and a gateway machine external to the cluster from which clients submit jobs. What do you need to do in order to run Impala on the cluster and submit jobs from the command line of the gateway machine?

    A. Install the impalad daemon statestored daemon, and daemon on each machine in the cluster, and the impala shell on your gateway machine

    B. Install the impalad daemon, the statestored daemon, the catalogd daemon, and the impala shell on your gateway machine

    C. Install the impalad daemon and the impala shell on your gateway machine, and the statestored daemon and catalogd daemon on one of the nodes in the cluster

    D. Install the impalad daemon on each machine in the cluster, the statestored daemon and catalogd daemon on one machine in the cluster, and the impala shell on your gateway machine

    E. Install the impalad daemon, statestored daemon, and catalogd daemon on each machine in the cluster and on the gateway node

  • Question 12:

    Your cluster implements HDFS High Availability (HA). Your two NameNodes are named nn01 and nn02. What occurs when you execute the command: hdfs haadmin failover nn01 nn02?

    A. nn02 is fenced, and nn01 becomes the active NameNode

    B. nn01 is fenced, and nn02 becomes the active NameNode

    C. nn01 becomes the standby NameNode and nn02 becomes the active NameNode

    D. nn02 becomes the standby NameNode and nn01 becomes the active NameNode

  • Question 13:

    Choose three reasons why should you run the HDFS balancer periodically? (Choose three)

    A. To ensure that there is capacity in HDFS for additional data

    B. To ensure that all blocks in the cluster are 128MB in size

    C. To help HDFS deliver consistent performance under heavy loads

    D. To ensure that there is consistent disk utilization across the DataNodes

    E. To improve data locality MapReduce

  • Question 14:

    You decide to create a cluster which runs HDFS in High Availability mode with automatic failover, using Quorum Storage. What is the purpose of ZooKeeper in such a configuration?

    A. It only keeps track of which NameNode is Active at any given time

    B. It monitors an NFS mount point and reports if the mount point disappears

    C. It both keeps track of which NameNode is Active at any given time, and manages the Edits file. Which is a log of changes to the HDFS filesystem

    D. If only manages the Edits file, which is log of changes to the HDFS filesystem

    E. Clients connect to ZooKeeper to determine which NameNode is Active

  • Question 15:

    You are running Hadoop cluster with all monitoring facilities properly configured.

    Which scenario will go undeselected?

    A. HDFS is almost full

    B. The NameNode goes down

    C. A DataNode is disconnected from the cluster

    D. Map or reduce tasks that are stuck in an infinite loop

    E. MapReduce jobs are causing excessive memory swaps

  • Question 16:

    You have A 20 node Hadoop cluster, with 18 slave nodes and 2 master nodes running HDFS High Availability (HA). You want to minimize the chance of data loss in your cluster. What should you do?

    A. Add another master node to increase the number of nodes running the JournalNode which increases the number of machines available to HA to create a quorum

    B. Set an HDFS replication factor that provides data redundancy, protecting against node failure

    C. Run a Secondary NameNode on a different master from the NameNode in order to provide automatic recovery from a NameNode failure.

    D. Run the ResourceManager on a different master from the NameNode in order to load-share HDFS metadata processing

    E. Configure the cluster's disk drives with an appropriate fault tolerant RAID level

  • Question 17:

    Your Hadoop cluster is configuring with HDFS and MapReduce version 2 (MRv2) on YARN. Can you configure a worker node to run a NodeManager daemon but not a DataNode daemon and still have a functional cluster?

    A. Yes. The daemon will receive data from the NameNode to run Map tasks

    B. Yes. The daemon will get data from another (non-local) DataNode to run Map tasks

    C. Yes. The daemon will receive Map tasks only

    D. Yes. The daemon will receive Reducer tasks only

  • Question 18:

    Your cluster has the following characteristics:

    A rack aware topology is configured and on

    Replication is set to 3

    Cluster block size is set to 64MB

    Which describes the file read process when a client application connects into the cluster and requests a 50MB file?

    A. The client queries the NameNode for the locations of the block, and reads all three copies. The first copy to complete transfer to the client is the one the client reads as part of hadoop's speculative execution framework.

    B. The client queries the NameNode for the locations of the block, and reads from the first location in the list it receives.

    C. The client queries the NameNode for the locations of the block, and reads from a random location in the list it receives to eliminate network I/O loads by balancing which nodes it retrieves data from any given time.

    D. The client queries the NameNode which retrieves the block from the nearest DataNode to the client then passes that block back to the client.

  • Question 19:

    You're upgrading a Hadoop cluster from HDFS and MapReduce version 1 (MRv1) to one running HDFS and MapReduce version 2 (MRv2) on YARN. You want to set and enforce version 1 (MRv1) to one running HDFS and MapReduce version 2 (MRv2) on YARN. You want to set and enforce a block size of 128MB for all new files written to the cluster after upgrade. What should you do?

    A. You cannot enforce this, since client code can always override this value

    B. Set dfs.block.size to 128M on all the worker nodes, on all client machines, and on the NameNode, and set the parameter to final

    C. Set dfs.block.size to 128 M on all the worker nodes and client machines, and set the parameter to final. You do not need to set this value on the NameNode

    D. Set dfs.block.size to 134217728 on all the worker nodes, on all client machines, and on the NameNode, and set the parameter to final

    E. Set dfs.block.size to 134217728 on all the worker nodes and client machines, and set the parameter to final. You do not need to set this value on the NameNode

  • Question 20:

    On a cluster running MapReduce v2 (MRv2) on YARN, a MapReduce job is given a directory of 10 plain text files as its input directory. Each file is made up of 3 HDFS blocks. How many Mappers will run?

    A. We cannot say; the number of Mappers is determined by the ResourceManager

    B. We cannot say; the number of Mappers is determined by the developer

    C. 30

    D. 3

    E. 10

    F. We cannot say; the number of mappers is determined by the ApplicationMaster

Tips on How to Prepare for the Exams

Nowadays, the certification exams become more and more important and required by more and more enterprises when applying for a job. But how to prepare for the exam effectively? How to prepare for the exam in a short time with less efforts? How to get a ideal result and how to find the most reliable resources? Here on Vcedump.com, you will find all the answers. Vcedump.com provide not only Cloudera exam questions, answers and explanations but also complete assistance on your exam preparation and certification application. If you are confused on your CCA-500 exam preparations and Cloudera certification application, do not hesitate to visit our Vcedump.com to find your solutions here.