Free Professional Machine Learning Engineer Exam Braindumps (page: 12)

Page 12 of 69

You are training a TensorFlow model on a structured dataset with 100 billion records stored in several CSV files. You need to improve the input/output execution performance. What should you do?

  1. Load the data into BigQuery, and read the data from BigQuery.
  2. Load the data into Cloud Bigtable, and read the data from Bigtable.
  3. Convert the CSV files into shards of TFRecords, and store the data in Cloud Storage.
  4. Convert the CSV files into shards of TFRecords, and store the data in the Hadoop Distributed File System (HDFS).

Answer(s): B


Reference:

https://cloud.google.com/dataflow/docs/guides/templates/provided-batch



As the lead ML Engineer for your company, you are responsible for building ML models to digitize scanned customer forms. You have developed a TensorFlow model that converts the scanned images into text and stores them in Cloud Storage. You need to use your ML model on the aggregated data collected at the end of each day with minimal manual intervention. What should you do?

  1. Use the batch prediction functionality of AI Platform.
  2. Create a serving pipeline in Compute Engine for prediction.
  3. Use Cloud Functions for prediction each time a new data point is ingested.
  4. Deploy the model on AI Platform and create a version of it for online inference.

Answer(s): A



You recently joined an enterprise-scale company that has thousands of datasets. You know that there are accurate descriptions for each table in BigQuery, and you are searching for the proper BigQuery table to use for a model you are building on AI Platform. How should you find the data that you need?

  1. Use Data Catalog to search the BigQuery datasets by using keywords in the table description.
  2. Tag each of your model and version resources on AI Platform with the name of the BigQuery table that was used for training.
  3. Maintain a lookup table in BigQuery that maps the table descriptions to the table ID. Query the lookup table to find the correct table ID for the data that you need.
  4. Execute a query in BigQuery to retrieve all the existing table names in your project using the INFORMATION_SCHEMA metadata tables that are native to BigQuery. Use the result o find the table that you need.

Answer(s): A



You started working on a classification problem with time series data and achieved an area under the receiver operating characteristic curve (AUC ROC) value of 99% for training data after just a few experiments. You haven’t explored using any sophisticated algorithms or spent any time on hyperparameter tuning. What should your next step be to identify and fix the problem?

  1. Address the model overfitting by using a less complex algorithm.
  2. Address data leakage by applying nested cross-validation during model training.
  3. Address data leakage by removing features highly correlated with the target value.
  4. Address the model overfitting by tuning the hyperparameters to reduce the AUC ROC value.

Answer(s): B



Page 12 of 69



Post your Comments and Discuss Google Professional Machine Learning Engineer exam with other Community members:

Tina commented on April 09, 2024
Good questions
Anonymous
upvote

Kavah commented on September 29, 2021
Very responsive and cool support team.
UNITED KINGDOM
upvote