Exam Details

  • Exam Code
    :PROFESSIONAL-MACHINE-LEARNING-ENGINEER
  • Exam Name
    :Professional Machine Learning Engineer
  • Certification
    :Google Certifications
  • Vendor
    :Google
  • Total Questions
    :282 Q&As
  • Last Updated
    :Apr 21, 2024

Google Google Certifications PROFESSIONAL-MACHINE-LEARNING-ENGINEER Questions & Answers

  • Question 1:

    Your team is building a convolutional neural network (CNN)-based architecture from scratch. The preliminary experiments running on your on-premises CPU-only infrastructure were encouraging, but have slow convergence. You have been asked to speed up model training to reduce time-to-market. You want to experiment with virtual machines (VMs) on Google Cloud to leverage more powerful hardware. Your code does not include any manual device placement and has not been wrapped in Estimator model-level abstraction. Which environment should you train your model on?

    A. AVM on Compute Engine and 1 TPU with all dependencies installed manually.

    B. AVM on Compute Engine and 8 GPUs with all dependencies installed manually.

    C. A Deep Learning VM with an n1-standard-2 machine and 1 GPU with all libraries pre-installed.

    D. A Deep Learning VM with more powerful CPU e2-highcpu-16 machines with all libraries pre-installed.

  • Question 2:

    You work for an online travel agency that also sells advertising placements on its website to other companies. You have been asked to predict the most relevant web banner that a user should see next. Security is important to your company. The model latency requirements are 300ms@p99, the inventory is thousands of web banners, and your exploratory analysis has shown that navigation context is a good predictor. You want to Implement the simplest solution. How should you configure the prediction pipeline?

    A. Embed the client on the website, and then deploy the model on AI Platform Prediction.

    B. Embed the client on the website, deploy the gateway on App Engine, and then deploy the model on AI Platform Prediction.

    C. Embed the client on the website, deploy the gateway on App Engine, deploy the database on Cloud Bigtable for writing and for reading the user's navigation context, and then deploy the model on AI Platform Prediction.

    D. Embed the client on the website, deploy the gateway on App Engine, deploy the database on Memorystore for writing and for reading the user's navigation context, and then deploy the model on Google Kubernetes Engine.

  • Question 3:

    You started working on a classification problem with time series data and achieved an area under the receiver operating characteristic curve (AUC ROC) value of 99% for training data after just a few experiments. You haven't explored using any sophisticated algorithms or spent any time on hyperparameter tuning. What should your next step be to identify and fix the problem?

    A. Address the model overfitting by using a less complex algorithm.

    B. Address data leakage by applying nested cross-validation during model training.

    C. Address data leakage by removing features highly correlated with the target value.

    D. Address the model overfitting by tuning the hyperparameters to reduce the AUC ROC value.

  • Question 4:

    You recently joined an enterprise-scale company that has thousands of datasets. You know that there are accurate descriptions for each table in BigQuery, and you are searching for the proper BigQuery table to use for a model you are building on AI Platform. How should you find the data that you need?

    A. Use Data Catalog to search the BigQuery datasets by using keywords in the table description.

    B. Tag each of your model and version resources on AI Platform with the name of the BigQuery table that was used for training.

    C. Maintain a lookup table in BigQuery that maps the table descriptions to the table ID. Query the lookup table to find the correct table ID for the data that you need.

    D. Execute a query in BigQuery to retrieve all the existing table names in your project using the INFORMATION_SCHEMA metadata tables that are native to BigQuery. Use the result o find the table that you need.

  • Question 5:

    As the lead ML Engineer for your company, you are responsible for building ML models to digitize scanned customer forms. You have developed a TensorFlow model that converts the scanned images into text and stores them in Cloud Storage. You need to use your ML model on the aggregated data collected at the end of each day with minimal manual intervention. What should you do?

    A. Use the batch prediction functionality of AI Platform.

    B. Create a serving pipeline in Compute Engine for prediction.

    C. Use Cloud Functions for prediction each time a new data point is ingested.

    D. Deploy the model on AI Platform and create a version of it for online inference.

  • Question 6:

    You are training a TensorFlow model on a structured dataset with 100 billion records stored in several CSV files. You need to improve the input/output execution performance. What should you do?

    A. Load the data into BigQuery, and read the data from BigQuery.

    B. Load the data into Cloud Bigtable, and read the data from Bigtable.

    C. Convert the CSV files into shards of TFRecords, and store the data in Cloud Storage.

    D. Convert the CSV files into shards of TFRecords, and store the data in the Hadoop Distributed File System (HDFS).

  • Question 7:

    You are an ML engineer at a global car manufacture. You need to build an ML model to predict car sales in different cities around the world. Which features or feature crosses should you use to train city-specific relationships between car type and number of sales?

    A. Thee individual features: binned latitude, binned longitude, and one-hot encoded car type.

    B. One feature obtained as an element-wise product between latitude, longitude, and car type.

    C. One feature obtained as an element-wise product between binned latitude, binned longitude, and one-hot encoded car type.

    D. Two feature crosses as an element-wise product: the first between binned latitude and one-hot encoded car type, and the second between binned longitude and one-hot encoded car type.

  • Question 8:

    You work for a large technology company that wants to modernize their contact center. You have been asked to develop a solution to classify incoming calls by product so that requests can be more quickly routed to the correct support team. You have already transcribed the calls using the Speech-to-Text API. You want to minimize data preprocessing and development time. How should you build the model?

    A. Use the AI Platform Training built-in algorithms to create a custom model.

    B. Use AutoMlL Natural Language to extract custom entities for classification.

    C. Use the Cloud Natural Language API to extract custom entities for classification.

    D. Build a custom model to identify the product keywords from the transcribed calls, and then run the keywords through a classification algorithm.

  • Question 9:

    You work for an advertising company and want to understand the effectiveness of your company's latest advertising campaign. You have streamed 500 MB of campaign data into BigQuery. You want to query the table, and then manipulate the results of that query with a pandas dataframe in an AI Platform notebook. What should you do?

    A. Use AI Platform Notebooks' BigQuery cell magic to query the data, and ingest the results as a pandas dataframe.

    B. Export your table as a CSV file from BigQuery to Google Drive, and use the Google Drive API to ingest the file into your notebook instance.

    C. Download your table from BigQuery as a local CSV file, and upload it to your AI Platform notebook instance. Use pandas.read_csv to ingest he file as a pandas dataframe.

    D. From a bash cell in your AI Platform notebook, use the bq extract command to export the table as a CSV file to Cloud Storage, and then use gsutil cp to copy the data into the notebook. Use pandas.read_csv to ingest the file as a pandas dataframe.

  • Question 10:

    You have a functioning end-to-end ML pipeline that involves tuning the hyperparameters of your ML model using AI Platform, and then using the best-tuned parameters for training. Hypertuning is taking longer than expected and is delaying the downstream processes. You want to speed up the tuning job without significantly compromising its effectiveness. Which actions should you take? (Choose two.)

    A. Decrease the number of parallel trials.

    B. Decrease the range of floating-point values.

    C. Set the early stopping parameter to TRUE.

    D. Change the search algorithm from Bayesian search to random search.

    E. Decrease the maximum number of trials during subsequent training phases.

Tips on How to Prepare for the Exams

Nowadays, the certification exams become more and more important and required by more and more enterprises when applying for a job. But how to prepare for the exam effectively? How to prepare for the exam in a short time with less efforts? How to get a ideal result and how to find the most reliable resources? Here on Vcedump.com, you will find all the answers. Vcedump.com provide not only Google exam questions, answers and explanations but also complete assistance on your exam preparation and certification application. If you are confused on your PROFESSIONAL-MACHINE-LEARNING-ENGINEER exam preparations and Google certification application, do not hesitate to visit our Vcedump.com to find your solutions here.