Exam Details

  • Exam Code
    :DATABRICKS-CERTIFIED-ASSOCIATE-DEVELOPER-FOR-APACHE-SPARK
  • Exam Name
    :Databricks Certified Associate Developer for Apache Spark 3.0
  • Certification
    :Databricks Certifications
  • Vendor
    :Databricks
  • Total Questions
    :180 Q&As
  • Last Updated
    :Jul 02, 2025

Databricks Databricks Certifications DATABRICKS-CERTIFIED-ASSOCIATE-DEVELOPER-FOR-APACHE-SPARK Questions & Answers

  • Question 51:

    Which of the following code blocks returns a single-column DataFrame of all entries in Python list throughputRates which contains only float-type values ?

    A. spark.createDataFrame((throughputRates), FloatType)

    B. spark.createDataFrame(throughputRates, FloatType)

    C. spark.DataFrame(throughputRates, FloatType)

    D. spark.createDataFrame(throughputRates)

    E. spark.createDataFrame(throughputRates, FloatType())

  • Question 52:

    Which of the following code blocks returns only rows from DataFrame transactionsDf in which values in column productId are unique?

    A. transactionsDf.distinct("productId")

    B. transactionsDf.dropDuplicates(subset=["productId"])

    C. transactionsDf.drop_duplicates(subset="productId")

    D. transactionsDf.unique("productId")

    E. transactionsDf.dropDuplicates(subset="productId")

  • Question 53:

    Which of the following describes the role of the cluster manager?

    A. The cluster manager schedules tasks on the cluster in client mode.

    B. The cluster manager schedules tasks on the cluster in local mode.

    C. The cluster manager allocates resources to Spark applications and maintains the executor processes in client mode.

    D. The cluster manager allocates resources to Spark applications and maintains the executor processes in remote mode.

    E. The cluster manager allocates resources to the DataFrame manager.

  • Question 54:

    Which of the following code blocks creates a new 6-column DataFrame by appending the rows of the 6column DataFrame yesterdayTransactionsDf to the rows of the 6-column DataFrame todayTransactionsDf, ignoring that both DataFrames have different column names?

    A. union(todayTransactionsDf, yesterdayTransactionsDf)

    B. todayTransactionsDf.unionByName(yesterdayTransactionsDf, allowMissingColumns=True)

    C. todayTransactionsDf.unionByName(yesterdayTransactionsDf)

    D. todayTransactionsDf.concat(yesterdayTransactionsDf)

    E. todayTransactionsDf.union(yesterdayTransactionsDf)

  • Question 55:

    Which of the following code blocks returns about 150 randomly selected rows from the 1000-row DataFrame transactionsDf, assuming that any row can appear more than once in the returned DataFrame?

    A. transactionsDf.resample(0.15, False, 3142)

    B. transactionsDf.sample(0.15, False, 3142)

    C. transactionsDf.sample(0.15)

    D. transactionsDf.sample(0.85, 8429)

    E. transactionsDf.sample(True, 0.15, 8261)

  • Question 56:

    Which of the following code blocks displays various aggregated statistics of all columns in DataFrame transactionsDf, including the standard deviation and minimum of values in each column?

    A. transactionsDf.summary()

    B. transactionsDf.agg("count", "mean", "stddev", "25%", "50%", "75%", "min")

    C. transactionsDf.summary("count", "mean", "stddev", "25%", "50%", "75%", "max").show()

    D. transactionsDf.agg("count", "mean", "stddev", "25%", "50%", "75%", "min").show()

    E. transactionsDf.summary().show()

  • Question 57:

    The code block displayed below contains an error. The code block should trigger Spark to cache DataFrame transactionsDf in executor memory where available, writing to disk where insufficient

    executor memory is available, in a fault-tolerant way. Find the error.

    Code block:

    transactionsDf.persist(StorageLevel.MEMORY_AND_DISK)

    A. Caching is not supported in Spark, data are always recomputed.

    B. Data caching capabilities can be accessed through the spark object, but not through the DataFrame API.

    C. The storage level is inappropriate for fault-tolerant storage.

    D. The code block uses the wrong operator for caching.

    E. The DataFrameWriter needs to be invoked.

  • Question 58:

    Which of the following describes Spark's way of managing memory?

    A. Spark uses a subset of the reserved system memory.

    B. Storage memory is used for caching partitions derived from DataFrames.

    C. As a general rule for garbage collection, Spark performs better on many small objects than few big objects.

    D. Disabling serialization potentially greatly reduces the memory footprint of a Spark application.

    E. Spark's memory usage can be divided into three categories: Execution, transaction, and storage.

  • Question 59:

    Which of the following describes the role of tasks in the Spark execution hierarchy?

    A. Tasks are the smallest element in the execution hierarchy.

    B. Within one task, the slots are the unit of work done for each partition of the data.

    C. Tasks are the second-smallest element in the execution hierarchy.

    D. Stages with narrow dependencies can be grouped into one task.

    E. Tasks with wide dependencies can be grouped into one stage.

  • Question 60:

    Which of the following describes properties of a shuffle?

    A. Operations involving shuffles are never evaluated lazily.

    B. Shuffles involve only single partitions.

    C. Shuffles belong to a class known as "full transformations".

    D. A shuffle is one of many actions in Spark.

    E. In a shuffle, Spark writes data to disk.

Tips on How to Prepare for the Exams

Nowadays, the certification exams become more and more important and required by more and more enterprises when applying for a job. But how to prepare for the exam effectively? How to prepare for the exam in a short time with less efforts? How to get a ideal result and how to find the most reliable resources? Here on Vcedump.com, you will find all the answers. Vcedump.com provide not only Databricks exam questions, answers and explanations but also complete assistance on your exam preparation and certification application. If you are confused on your DATABRICKS-CERTIFIED-ASSOCIATE-DEVELOPER-FOR-APACHE-SPARK exam preparations and Databricks certification application, do not hesitate to visit our Vcedump.com to find your solutions here.