DAS-C01 Exam Details

  • Exam Code
    :DAS-C01
  • Exam Name
    :AWS Certified Data Analytics - Specialty (DAS-C01)
  • Certification
    :Amazon Certifications
  • Vendor
    :Amazon
  • Total Questions
    :285 Q&As
  • Last Updated
    :Jan 12, 2026

Amazon DAS-C01 Online Questions & Answers

  • Question 1:

    A retail company uses Amazon Aurora MySQL for its operational data store and Amazon Redshift for its data warehouse. The MySQL database resides in a different VPC than the Redshift cluster. Data analysts need to query data in both MySQL and Amazon Redshift to provide business insights. The company wants the lowest network latency between the two VPCs.

    Which combination of solutions meet these requirements? (Choose two.)

    A. Set up VPC peering between the MySQL VPC and the Redshift VPC.
    B. Set up a transit gateway to connect the MySQL VPC with the Redshift VPC.
    C. Use a Redshift federated query to retrieve live data from the MySQL database. Create a late-binding view to combine the MySQL data with the Redshift data.
    D. Use Amazon Redshift Spectrum to retrieve live data from the MySQL database. Create a late-binding view to combine the MySQL data with the Redshift data.
    E. Use the Redshift COPY command to constantly copy live data from MySQL to the Redshift cluster. Create a late-binding view to combine the MySQL data with the Redshift data.

  • Question 2:

    An online gaming company wants to read customer data from Amazon Kinesis Data Streams and deliver the data to an Amazon S3 data lake for analytics. The data contains customer_id as one of the attributes. The data consumers need the data to be partitioned by customer_id in the S3 data lake.

    Which solution will meet this requirement with the LEAST effort?

    A. Create an Amazon Kinesis Data Firehose delivery stream. Use dynamic partitioning to partition the data by customer_id before delivering the data to Amazon S3.
    B. Create an AWS Glue streaming job Use the built-in map transform to partition the data by customer_id before delivering the data to Amazon S3.
    C. Create an AWS Lambda function. Use Lambda layers to partition the data by customer_id before delivering the data to Amazon S3.
    D. Create an Amazon EMR cluster. Run an Apache Spark job to automatically partition the data by customer_id before delivering the data to Amazon S3.

  • Question 3:

    A mining company is using Amazon S3 as its data lake. The company wants to analyze the data collected by the sensors in its mines. A data pipeline is being built to capture data from the sensors, ingest the data into an S3 bucket, and convert the data to Apache Parquet format. The data pipeline must be processed in near-real time. The data will be used for on-demand queries with Amazon Athena.

    Which solution will meet these requirements?

    A. Use Amazon Kinesis Data Firehose to invoke an AWS Lambda function that converts the data to Parquet format and stores the data in Amazon S3.
    B. Use Amazon Kinesis Data Streams to invoke an AWS Lambda function that converts the data to Parquet format and stores the data in Amazon S3.
    C. Use AWS DataSync to invoke an AWS Lambda function that converts the data to Parquet format and stores the data in Amazon S3.
    D. Use Amazon Simple Queue Service (Amazon SQS) to stream data directly to an AWS Glue job that converts the data to Parquet format and stores the data in Amazon S3.

  • Question 4:

    A company ingests a large set of sensor data in nested JSON format from different sources and stores it in an Amazon S3 bucket. The sensor data must be joined with performance data currently stored in an Amazon Redshift cluster.

    A business analyst with basic SQL skills must build dashboards and analyze this data in Amazon QuickSight. A data engineer needs to build a solution to prepare the data for use by the business analyst. The data engineer does not know the

    structure of the JSON file. The company requires a solution with the least possible implementation effort.

    Which combination of steps will create a solution that meets these requirements? (Choose three.)

    A. Use an AWS Glue ETL job to convert the data into Apache Parquet format and write to Amazon S3.
    B. Use an AWS Glue crawler to catalog the data.
    C. Use an AWS Glue ETL job with the ApplyMapping class to un-nest the data and write to Amazon Redshift tables.
    D. Use an AWS Glue ETL job with the Regionalize class to un-nest the data and write to Amazon Redshift tables.
    E. Use QuickSight to create an Amazon Athena data source to read the Apache Parquet files in Amazon S3.
    F. Use QuickSight to create an Amazon Redshift data source to read the native Amazon Redshift tables.

  • Question 5:

    A company has an electronic healthcare system that contains data of patients. The data is consolidated from multiple systems and is stored in an Amazon S3 bucket in .csv format. The company has created an AWS Glue Data Catalog. The

    dataset contains duplicate data, and no unique keys exist to identify a patient. Fields do not match exactly across the systems.

    A data analytics specialist must design a solution to identify and remove duplicates. The solution must minimize the amount of human intervention and code that are required.

    The data analytics specialist starts by using labeled data to teach the FindMatches machine learning (ML) transform.

    What must the data analytics specialist do next to meet these requirements?

    A. Identify matches in the dataset by using an AWS Glue ETL job with Spark distinct(). Review the output by using Amazon Redshift Spectrum.
    B. Identify matches in the dataset by using an AWS Glue ETL job with Spark distinct(). Create a Data Catalog of transformed results Review the output by using Amazon Athena.
    C. Identify matches in the dataset by using an AWS Glue ETL job that has a transform type of "find matching records." Create a Data Catalog of transformed results. Review the output by using Amazon Athena.
    D. Identify matches in the dataset by using an AWS Glue ETL job that has a transform type of "find matching records." Review the output by using Amazon Redshift Spectrum.

  • Question 6:

    A company uses an Amazon QuickSight business intelligence (BI) dashboard to make decisions. A data warehouse that is hosted on an Amazon Redshift cluster is the primary data source for the dashboard.

    The user experience has degraded because some BI queries are complex and include multiple table joins. A data analytics specialist needs to reduce query response times to improve the user experience and meet SLAs.

    Which solution will meet these requirements with the LEAST operational overhead?

    A. Create automated materialized views. Turn on the query rewrite feature in Amazon Redshift.
    B. Create views in Amazon Redshift. Change the QuickSight code to select from the views.
    C. Create a new consolidated table by using CREATE TABLE AS (CTAS). Change the QuickSight code to select from the consolidated table.
    D. Create an external table. Turn on the query rewrite feature in Amazon Redshift.

  • Question 7:

    A financial services company hosts its data warehouse on a Hadoop cluster located in an on-premises data center. The data is 300 TB in size and grows by 1 TB each day. The data is generated in real time from the company's trading system. The raw data is transformed at the end of the trading day using a custom tool running on the Hadoop cluster.

    The company is migrating its data warehouse to AWS using a managed data warehouse product provided by a third party that can ingest data from Amazon S3. The company has already established a 10 Gbps connection with an AWS Region using AWS Direct Connect. The company is required by its security and regulatory compliance policies to not transfer data over the public internet. The company wants to minimize changes to its custom tool for data transformation. The company also plans to eliminate the on-premises Hadoop cluster after the migration.

    Which solution MOST cost-effectively meets these requirements?

    A. Create a VPC endpoint for Amazon S3. Run a one-time copy job using the DistCp tool to copy existing files from Hadoop to a target S3 bucket over the VPC endpoint Schedule a nightly DistCp job on the Hadoop cluster to copy the incremental files produced by the custom tool to the target S3 bucket
    B. Create a VPC endpoint for Amazon S3. Run a one-time copy job using the DistCp tool to copy existing files from Hadoop to a target S3 bucket over the VPC endpoint. Schedule a nightly job on the trading system servers that produces raw data to copy the incremental raw files to the target S3 bucket. Run the data transformation tool on a transient Amazon EMR cluster to output files to Amazon S3.
    C. Create a VPC endpoint for Amazon S3. Run a one-time copy job using the DistCp tool to copy existing files from Hadoop to a target S3 bucket over the VPC endpoint. Set up an Amazon Kinesis data stream to ingest raw data from the trading system in real time. Use Amazon Kinesis Data Analytics to transform the raw data and output files to Amazon S3.
    D. Complete a one-time transfer of the data using AWS Snowball Edge devices transferring to a target S3 bucket. Schedule a nightly job on the trading system servers that produces raw data to copy the incremental raw files to the target S3 bucket Run the data transformation tool on a transient Amazon EMR cluster to output files to Amazon S3.

  • Question 8:

    A company is attempting to use Amazon Athena to run analytics queries. The company is using the UNLOAD statement and has noticed that the result of the query is not appearing in the destination Amazon S3 bucket. The query previously ran successfully, and the company has not changed any configuration since then.

    Which reason would prevent the query from running successfully?

    A. The destination S3 bucket does not have the output folder.
    B. The destination S3 bucket contains files from previous queries.
    C. The destination S3 bucket has an S3 Lifecycle policy configured.
    D. The destination S3 bucket has versioning enabled.

  • Question 9:

    A company's system operators and security engineers need to analyze activities within specific date ranges of AWS CloudTrail logs. All log files are stored in an Amazon S3 bucket, and the size of the logs is more than 5 TB. The solution must be cost-effective and maximize query performance.

    Which solution meets these requirements?

    A. Copy the logs to a new S3 bucket with a prefix structure of . Use the date column as a partition key. Create a table on Amazon Athena based on the objects in the new bucket. Automatically add metadata partitions by using the MSCK REPAIR TABLE command in Athena. Use Athena to query the table and partitions.
    B. Create a table on Amazon Athena. Manually add metadata partitions by using the ALTER TABLE ADD PARTITION statement, and use multiple columns for the partition key. Use Athena to query the table and partitions.
    C. Launch an Amazon EMR cluster and use Amazon S3 as a data store for Apache HBase. Load the logs from the S3 bucket to an HBase table on Amazon EMR. Use Amazon Athena to query the table and partitions.
    D. Create an AWS Glue job to copy the logs from the S3 source bucket to a new S3 bucket and create a table using Apache Parquet file format, Snappy as compression codec, and partition by date. Use Amazon Athena to query the table and partitions.

  • Question 10:

    An IoT company is collecting data from multiple sensors and is streaming the data to Amazon Managed Streaming for Apache Kafka (Amazon MSK). Each sensor type has its own topic, and each topic has the same number of partitions.

    The company is planning to turn on more sensors. However, the company wants to evaluate which sensor types are producing the most data so that the company can scale accordingly. The company needs to know which sensor types have

    the largest values for the following metrics: BytesInPerSec and MessagesInPerSec.

    Which level of monitoring for Amazon MSK will meet these requirements?

    A. DEFAULT level
    B. PER_TOPIC_PER_BROKER level
    C. PER_BROKER level
    D. PER_TOPIC level

Tips on How to Prepare for the Exams

Nowadays, the certification exams become more and more important and required by more and more enterprises when applying for a job. But how to prepare for the exam effectively? How to prepare for the exam in a short time with less efforts? How to get a ideal result and how to find the most reliable resources? Here on Vcedump.com, you will find all the answers. Vcedump.com provide not only Amazon exam questions, answers and explanations but also complete assistance on your exam preparation and certification application. If you are confused on your DAS-C01 exam preparations and Amazon certification application, do not hesitate to visit our Vcedump.com to find your solutions here.