Exam Details

  • Exam Code
    :CCA-500
  • Exam Name
    :Cloudera Certified Administrator for Apache Hadoop (CCAH)
  • Certification
    :CCAH
  • Vendor
    :Cloudera
  • Total Questions
    :60 Q&As
  • Last Updated
    :May 14, 2024

Cloudera CCAH CCA-500 Questions & Answers

  • Question 41:

    Which scheduler would you deploy to ensure that your cluster allows short jobs to finish within a reasonable time without starting long-running jobs?

    A. Complexity Fair Scheduler (CFS)

    B. Capacity Scheduler

    C. Fair Scheduler

    D. FIFO Scheduler

  • Question 42:

    Your cluster is configured with HDFS and MapReduce version 2 (MRv2) on YARN. What is the result when you execute: hadoop jar SampleJar MyClass on a client machine?

    A. SampleJar.Jar is sent to the ApplicationMaster which allocates a container for SampleJar.Jar

    B. Sample.jar is placed in a temporary directory in HDFS

    C. SampleJar.jar is sent directly to the ResourceManager

    D. SampleJar.jar is serialized into an XML file which is submitted to the ApplicatoionMaster

  • Question 43:

    Which YARN process run as "container 0" of a submitted job and is responsible for resource qrequests?

    A. ApplicationManager

    B. JobTracker

    C. ApplicationMaster

    D. JobHistoryServer

    E. ResoureManager

    F. NodeManager

  • Question 44:

    Which YARN daemon or service monitors a Controller's per-application resource using (e.g., memory CPU)?

    A. ApplicationMaster

    B. NodeManager

    C. ApplicationManagerService

    D. ResourceManager

  • Question 45:

    Which is the default scheduler in YARN?

    A. YARN doesn't configure a default scheduler, you must first assign an appropriate scheduler class in yarn-site.xml

    B. Capacity Scheduler

    C. Fair Scheduler

    D. FIFO Scheduler

  • Question 46:

    Identify two features/issues that YARN is designated to address: (Choose two)

    A. Standardize on a single MapReduce API

    B. Single point of failure in the NameNode

    C. Reduce complexity of the MapReduce APIs

    D. Resource pressure on the JobTracker

    E. Ability to run framework other than MapReduce, such as MPI

    F. HDFS latency

  • Question 47:

    You need to analyze 60,000,000 images stored in JPEG format, each of which is approximately 25 KB. Because you Hadoop cluster isn't optimized for storing and processing many small files, you decide to do the following actions:

    1.

    Group the individual images into a set of larger files

    2.

    Use the set of larger files as input for a MapReduce job that processes them directly with python using Hadoop streaming.

    Which data serialization system gives the flexibility to do this?

    A. CSV

    B. XML

    C. HTML

    D. Avro

    E. SequenceFiles

    F. JSON

  • Question 48:

    Which three basic configuration parameters must you set to migrate your cluster from MapReduce 1 (MRv1) to MapReduce V2 (MRv2)? (Choose three)

    A. Configure the NodeManager to enable MapReduce services on YARN by setting the following property in yarn-site.xml: yarn.nodemanager.hostname your_nodeManager_shuffle

    B. Configure the NodeManager hostname and enable node services on YARN by setting the following property in yarn-site.xml: yarn.nodemanager.hostname your_nodeManager_hostname

    C. Configure a default scheduler to run on YARN by setting the following property in mapred- site.xml: mapreduce.jobtracker.taskScheduler org.apache.hadoop.mapred.JobQueueTaskScheduler

    D. Configure the number of map tasks per jon YARN by setting the following property in mapred: mapreduce.job.maps 2

    E. Configure the ResourceManager hostname and enable node services on YARN by setting the following property in yarn-site.xml: yarn.resourcemanager.hostname your_resourceManager_hostname

    F. Configure MapReduce as a Framework running on YARN by setting the following property in mapredsite.xml: mapreduce.framework.name yarn

  • Question 49:

    What does CDH packaging do on install to facilitate Kerberos security setup?

    A. Automatically configures permissions for log files at and MAPRED_LOG_DIR/userlogs

    B. Creates users for hdfs and mapreduce to facilitate role assignment

    C. Creates directories for temp, hdfs, and mapreduce with the correct permissions

    D. Creates a set of pre-configured Kerberos keytab files and their permissions

    E. Creates and configures your kdc with default cluster values

  • Question 50:

    You want to understand more about how users browse your public website. For example, you want to know which pages they visit prior to placing an order. You have a server farm of 200 web servers hosting your website. Which is the most efficient process to gather these web server across logs into your Hadoop cluster analysis?

    A. Sample the web server logs web servers and copy them into HDFS using curl

    B. Ingest the server web logs into HDFS using Flume

    C. Channel these clickstreams into Hadoop using Hadoop Streaming

    D. Import all user clicks from your OLTP databases into Hadoop using Sqoop

    E. Write a MapReeeduce job with the web servers for mappers and the Hadoop cluster nodes for reducers

Tips on How to Prepare for the Exams

Nowadays, the certification exams become more and more important and required by more and more enterprises when applying for a job. But how to prepare for the exam effectively? How to prepare for the exam in a short time with less efforts? How to get a ideal result and how to find the most reliable resources? Here on Vcedump.com, you will find all the answers. Vcedump.com provide not only Cloudera exam questions, answers and explanations but also complete assistance on your exam preparation and certification application. If you are confused on your CCA-500 exam preparations and Cloudera certification application, do not hesitate to visit our Vcedump.com to find your solutions here.