Exam Details

  • Exam Code
    :PROFESSIONAL-DATA-ENGINEER
  • Exam Name
    :Professional Data Engineer on Google Cloud Platform
  • Certification
    :Google Certifications
  • Vendor
    :Google
  • Total Questions
    :331 Q&As
  • Last Updated
    :May 19, 2025

Google Google Certifications PROFESSIONAL-DATA-ENGINEER Questions & Answers

  • Question 181:

    Your team is working on a binary classification problem. You have trained a support vector machine (SVM) classifier with default parameters, and received an area under the Curve (AUC) of 0.87 on the validation set. You want to increase the AUC of the model. What should you do?

    A. Perform hyperparameter tuning

    B. Train a classifier with deep neural networks, because neural networks would always beat SVMs

    C. Deploy the model and measure the real-world AUC; it's always higher because of generalization

    D. Scale predictions you get out of the model (tune a scaling factor as a hyperparameter) in order to get the highest AUC

  • Question 182:

    You have data located in BigQuery that is used to generate reports for your company. You have noticed some weekly executive report fields do not correspond to format according to company standards for example, report errors include different telephone formats and different country code identifiers. This is a frequent issue, so you need to create a recurring job to normalize the data. You want a quick solution that requires no coding What should you do?

    A. Use Cloud Data Fusion and Wrangler to normalize the data, and set up a recurring job.

    B. Use BigQuery and GoogleSQL to normalize the data, and schedule recurring quenes in BigQuery.

    C. Create a Spark job and submit it to Dataproc Serverless.

    D. Use Dataflow SQL to create a job that normalizes the data, and that after the first run of the job, schedule the pipeline to execute recurrently.

  • Question 183:

    You are a head of BI at a large enterprise company with multiple business units that each have different priorities and budgets. You use on-demand pricing for BigQuery with a quota of 2K concurrent on-demand slots per project. Users at your organization sometimes don't get slots to execute their query and you need to correct this. You'd like to avoid introducing new projects to your account.

    What should you do?

    A. Convert your batch BQ queries into interactive BQ queries.

    B. Create an additional project to overcome the 2K on-demand per-project quota.

    C. Switch to flat-rate pricing and establish a hierarchical priority model for your projects.

    D. Increase the amount of concurrent slots per project at the Quotas page at the Cloud Console.

  • Question 184:

    Your financial services company is moving to cloud technology and wants to store 50 TB of financial timeseries data in the cloud. This data is updated frequently and new data will be streaming in all the time. Your company also wants to move their existing Apache Hadoop jobs to the cloud to get insights into this data.

    Which product should they use to store the data?

    A. Cloud Bigtable

    B. Google BigQuery

    C. Google Cloud Storage

    D. Google Cloud Datastore

  • Question 185:

    You have Cloud Functions written in Node.js that pull messages from Cloud Pub/Sub and send the data to BigQuery. You observe that the message processing rate on the Pub/Sub topic is orders of magnitude higher than anticipated, but there is no error logged in Stackdriver Log Viewer. What are the two most likely causes of this problem? Choose 2 answers.

    A. Publisher throughput quota is too small.

    B. Total outstanding messages exceed the 10-MB maximum.

    C. Error handling in the subscriber code is not handling run-time errors properly.

    D. The subscriber code cannot keep up with the messages.

    E. The subscriber code does not acknowledge the messages that it pulls.

  • Question 186:

    You need to look at BigQuery data from a specific table multiple times a day. The underlying table you are querying is several petabytes in size, but you want to filter your data and provide simple aggregations to downstream users. You want to run queries faster and get up-to-date insights quicker. What should you do?

    A. Run a scheduled query to pull the necessary data at specific intervals daily.

    B. Create a materialized view based off of the query being run.

    C. Use a cached query to accelerate time to results.

    D. Limit the query columns being pulled in the final result.

  • Question 187:

    Your company has a hybrid cloud initiative. You have a complex data pipeline that moves data between cloud provider services and leverages services from each of the cloud providers. Which cloud-native service should you use to orchestrate the entire pipeline?

    A. Cloud Dataflow

    B. Cloud Composer

    C. Cloud Dataprep

    D. Cloud Dataproc

  • Question 188:

    You are migrating your data warehouse to BigQuery. You have migrated all of your data into tables in a dataset. Multiple users from your organization will be using the data. They should only see certain tables based on their team membership. How should you set user permissions?

    A. Assign the users/groups data viewer access at the table level for each table

    B. Create SQL views for each team in the same dataset in which the data resides, and assign the users/groups data viewer access to the SQL views

    C. Create authorized views for each team in the same dataset in which the data resides, and assign the users/groups data viewer access to the authorized views

    D. Create authorized views for each team in datasets created for each team. Assign the authorized views data viewer access to the dataset in which the data resides. Assign the users/groups data viewer access to the datasets in which the authorized views reside

  • Question 189:

    You want to archive data in Cloud Storage. Because some data is very sensitive, you want to use the "Trust No One" (TNO) approach to encrypt your data to prevent the cloud provider staff from decrypting your data. What should you do?

    A. Use gcloud kms keys create to create a symmetric key. Then use gcloud kms encrypt to encrypt each archival file with the key and unique additional authenticated data (AAD). Use gsutil cp to upload each encrypted file to the Cloud Storage bucket, and keep the AAD outside of Google Cloud.

    B. Use gcloud kms keys create to create a symmetric key. Then use gcloud kms encrypt to encrypt each archival file with the key. Use gsutil cp to upload each encrypted file to the Cloud Storage bucket. Manually destroy the key previously used for encryption, and rotate the key once and rotate the key once.

    C. Specify customer-supplied encryption key (CSEK) in the .boto configuration file. Use gsutil cp to upload each archival file to the Cloud Storage bucket. Save the CSEK in Cloud Memorystore as permanent storage of the secret.

    D. Specify customer-supplied encryption key (CSEK) in the .boto configuration file. Use gsutil cp to upload each archival file to the Cloud Storage bucket. Save the CSEK in a different project that only the security team can access.

  • Question 190:

    You want to automate execution of a multi-step data pipeline running on Google Cloud. The pipeline includes Cloud Dataproc and Cloud Dataflow jobs that have multiple dependencies on each other. You want to use managed services where possible, and the pipeline will run every day. Which tool should you use?

    A. cron

    B. Cloud Composer

    C. Cloud Scheduler

    D. Workflow Templates on Cloud Dataproc

Tips on How to Prepare for the Exams

Nowadays, the certification exams become more and more important and required by more and more enterprises when applying for a job. But how to prepare for the exam effectively? How to prepare for the exam in a short time with less efforts? How to get a ideal result and how to find the most reliable resources? Here on Vcedump.com, you will find all the answers. Vcedump.com provide not only Google exam questions, answers and explanations but also complete assistance on your exam preparation and certification application. If you are confused on your PROFESSIONAL-DATA-ENGINEER exam preparations and Google certification application, do not hesitate to visit our Vcedump.com to find your solutions here.