Exam Details

  • Exam Code
    :CCA175
  • Exam Name
    :CCA Spark and Hadoop Developer Exam
  • Certification
    :Cloudera Certified Associate CCA
  • Vendor
    :Cloudera
  • Total Questions
    :95 Q&As
  • Last Updated
    :May 12, 2024

Cloudera Cloudera Certified Associate CCA CCA175 Questions & Answers

  • Question 81:

    Problem Scenario 52 : You have been given below code snippet.

    val b = sc.parallelize(List(1,2,3,4,5,6,7,8,2,4,2,1,1,1,1,1))

    Operation_xyz

    Write a correct code snippet for Operation_xyz which will produce below output.

    scalaxollection.Map[lnt,Long] = Map(5 -> 1, 8 -> 1, 3 -> 1, 6 -> 1, 1 -> S, 2 -> 3, 4 -> 2, 7 ->

    1)

  • Question 82:

    Problem Scenario 11 : You have been given following mysql database details as well as other info. user=retail_dba password=cloudera database=retail_db jdbc URL = jdbc:mysql://quickstart:3306/retail_db

    Please accomplish following.

    1.

    Import departments table in a directory called departments.

    2.

    Once import is done, please insert following 5 records in departments mysql table.

    Insert into departments(10, physics);

    Insert into departments(11, Chemistry);

    Insert into departments(12, Maths);

    Insert into departments(13, Science);

    Insert into departments(14, Engineering);

    3.

    Now import only new inserted records and append to existring directory . which has been created in first step.

  • Question 83:

    Problem Scenario 29 : Please accomplish the following exercises using HDFS command line options.

    1.

    Create a directory in hdfs named hdfs_commands.

    2.

    Create a file in hdfs named data.txt in hdfs_commands.

    3.

    Now copy this data.txt file on local filesystem, however while copying file please make sure file properties are not changed e.g. file permissions.

    4.

    Now create a file in local directory named data_local.txt and move this file to hdfs in hdfs_commands directory.

    5.

    Create a file data_hdfs.txt in hdfs_commands directory and copy it to local file system.

    6.

    Create a file in local filesystem named file1.txt and put it to hdfs

  • Question 84:

    Problem Scenario 85 : In Continuation of previous question, please accomplish following activities.

    1.

    Select all the columns from product table with output header as below. productID AS ID code AS Code name AS Description price AS 'Unit Price'

    2.

    Select code and name both separated by ' -' and header name should be Product Description'.

    3.

    Select all distinct prices.

    4.

    Select distinct price and name combination.

    5.

    Select all price data sorted by both code and productID combination.

    6.

    count number of products.

    7.

    Count number of products for each code.

  • Question 85:

    Problem Scenario 57 : You have been given below code snippet.

    val a = sc.parallelize(1 to 9, 3) operationl

    Write a correct code snippet for operationl which will produce desired output, shown below.

    Array[(String, Seq[lnt])] = Array((even,ArrayBuffer(2, 4, G, 8)), (odd,ArrayBuffer(1, 3, 5, 7,

    9)))

  • Question 86:

    Problem Scenario 24 : You have been given below comma separated employee information.

    Data Set:

    name,salary,sex,age alok,100000,male,29 jatin,105000,male,32 yogesh,134000,male,39 ragini,112000,female,35 jyotsana,129000,female,39 valmiki,123000,male,29

    Requirements:

    Use the netcat service on port 44444, and nc above data line by line. Please do the following activities.

    1.

    Create a flume conf file using fastest channel, which write data in hive warehouse directory, in a table called flumemaleemployee (Create hive table as well tor given data).

    2.

    While importing, make sure only male employee data is stored.

  • Question 87:

    Problem Scenario 54 : You have been given below code snippet.

    val a = sc.parallelize(List("dog", "tiger", "lion", "cat", "panther", "eagle"))

    val b = a.map(x => (x.length, x))

    operation1

    Write a correct code snippet for operationl which will produce desired output, shown below.

    Array[(lnt, String)] = Array((4,lion), (7,panther), (3,dogcat), (5,tigereagle))

  • Question 88:

    Problem Scenario 22 : You have been given below comma separated employee information. name,salary,sex,age alok,100000,male,29 jatin,105000,male,32 yogesh,134000,male,39 ragini,112000,female,35 jyotsana,129000,female,39 valmiki,123000,male,29 Use the netcat service on port 44444, and nc above data line by line. Please do the following activities.

    1.

    Create a flume conf file using fastest channel, which write data in hive warehouse directory, in a table called flumeemployee (Create hive table as well tor given data).

    2.

    Write a hive query to read average salary of all employees.

  • Question 89:

    Problem Scenario 18 : You have been given following mysql database details as well as other info. user=retail_dba password=cloudera database=retail_db jdbc URL = jdbc:mysql://quickstart:3306/retail_db Now accomplish following activities.

    1.

    Create mysql table as below.

    mysql --user=retail_dba -password=cloudera

    use retail_db

    CREATE TABLE IF NOT EXISTS departments_hive02(id int, department_name

    varchar(45), avg_salary int);

    show tables;

    2.

    Now export data from hive table departments_hive01 in departments_hive02. While

    exporting, please note following. wherever there is a empty string it should be loaded as a null value in

    mysql.

    wherever there is -999 value for int field, it should be created as null value.

  • Question 90:

    Problem Scenario 69 : Write down a Spark Application using Python, In which it read a file "Content.txt" (On hdfs) with following content. And filter out the word which is less than 2 characters and ignore all empty lines. Once doen store the filtered data in a directory called "problem84" (On hdfs) Content.txt Hello this is ABCTECH.com This is ABYTECH.com Apache Spark TrainingThis is Spark Learning Session Spark is faster than MapReduce

Tips on How to Prepare for the Exams

Nowadays, the certification exams become more and more important and required by more and more enterprises when applying for a job. But how to prepare for the exam effectively? How to prepare for the exam in a short time with less efforts? How to get a ideal result and how to find the most reliable resources? Here on Vcedump.com, you will find all the answers. Vcedump.com provide not only Cloudera exam questions, answers and explanations but also complete assistance on your exam preparation and certification application. If you are confused on your CCA175 exam preparations and Cloudera certification application, do not hesitate to visit our Vcedump.com to find your solutions here.