Free Professional Machine Learning Engineer Exam Braindumps (page: 38)

Page 38 of 69

You recently deployed an ML model. Three months after deployment, you notice that your model is underperforming on certain subgroups, thus potentially leading to biased results. You suspect that the inequitable performance is due to class imbalances in the training data, but you cannot collect more data. What should you do? (Choose two.)

  1. Remove training examples of high-performing subgroups, and retrain the model.
  2. Add an additional objective to penalize the model more for errors made on the minority class, and retrain the model
  3. Remove the features that have the highest correlations with the majority class.
  4. Upsample or reweight your existing training data, and retrain the model
  5. Redeploy the model, and provide a label explaining the model's behavior to users.

Answer(s): B,D



You are working on a binary classification ML algorithm that detects whether an image of a classified scanned document contains a company’s logo. In the dataset, 96% of examples don’t have the logo, so the dataset is very skewed. Which metric would give you the most confidence in your model?

  1. Precision
  2. Recall
  3. RMSE
  4. F1 score

Answer(s): D



While running a model training pipeline on Vertex Al, you discover that the evaluation step is failing because of an out-of-memory error. You are currently using TensorFlow Model Analysis (TFMA) with a standard Evaluator TensorFlow Extended (TFX) pipeline component for the evaluation step. You want to stabilize the pipeline without downgrading the evaluation quality while minimizing infrastructure overhead. What should you do?

  1. Include the flag -runner=DataflowRunner in beam_pipeline_args to run the evaluation step on Dataflow.
  2. Move the evaluation step out of your pipeline and run it on custom Compute Engine VMs with sufficient memory.
  3. Migrate your pipeline to Kubeflow hosted on Google Kubernetes Engine, and specify the appropriate node parameters for the evaluation step.
  4. Add tfma.MetricsSpec () to limit the number of metrics in the evaluation step.

Answer(s): C



You are developing an ML model using a dataset with categorical input variables. You have randomly split half of the data into training and test sets. After applying one-hot encoding on the categorical variables in the training set, you discover that one categorical variable is missing from the test set. What should you do?

  1. Use sparse representation in the test set.
  2. Randomly redistribute the data, with 70% for the training set and 30% for the test set
  3. Apply one-hot encoding on the categorical variables in the test data
  4. Collect more data representing all categories

Answer(s): B



Page 38 of 69



Post your Comments and Discuss Google Professional Machine Learning Engineer exam with other Community members:

Tina commented on April 09, 2024
Good questions
Anonymous
upvote

Kavah commented on September 29, 2021
Very responsive and cool support team.
UNITED KINGDOM
upvote