Free Professional Machine Learning Engineer Exam Braindumps (page: 14)

Page 14 of 69

Your team is working on an NLP research project to predict political affiliation of authors based on articles they have written. You have a large training dataset that is structured like this:


You followed the standard 80%-10%-10% data distribution across the training, testing, and evaluation subsets. How should you distribute the training examples across the train-test-eval subsets while maintaining the 80-10- 10 proportion?

  1. Distribute texts randomly across the train-test-eval subsets:
    Train set: [TextA1, TextB2, ...]
    Test set: [TextA2, TextC1, TextD2, ...]
    Eval set: [TextB1, TextC2, TextD1, ...]
  2. Distribute authors randomly across the train-test-eval subsets: (*)
    Train set: [TextA1, TextA2, TextD1, TextD2, ...]
    Test set: [TextB1, TextB2, ...]
    Eval set: [TexC1,TextC2 ...]
  3. Distribute sentences randomly across the train-test-eval subsets:
    Train set: [SentenceA11, SentenceA21, SentenceB11, SentenceB21, SentenceC11, SentenceD21 ...]
    Test set: [SentenceA12, SentenceA22, SentenceB12, SentenceC22, SentenceC12, SentenceD22 ...]
    Eval set: [SentenceA13, SentenceA23, SentenceB13, SentenceC23, SentenceC13, SentenceD31 ...]
  4. Distribute paragraphs of texts (i.e., chunks of consecutive sentences) across the train-test-eval subsets:
    Train set: [SentenceA11, SentenceA12, SentenceD11, SentenceD12 ...]
    Test set: [SentenceA13, SentenceB13, SentenceB21, SentenceD23, SentenceC12, SentenceD13 ...]
    Eval set: [SentenceA11, SentenceA22, SentenceB13, SentenceD22, SentenceC23, SentenceD11 ...]

Answer(s): B



Your team has been tasked with creating an ML solution in Google Cloud to classify support requests for one of your platforms. You analyzed the requirements and decided to use TensorFlow to build the classifier so that you have full control of the model’s code, serving, and deployment. You will use Kubeflow pipelines for the ML platform. To save time, you want to build on existing resources and use managed services instead of building a completely new model. How should you build the classifier?

  1. Use the Natural Language API to classify support requests.
  2. Use AutoML Natural Language to build the support requests classifier.
  3. Use an established text classification model on AI Platform to perform transfer learning.
  4. Use an established text classification model on AI Platform as-is to classify support requests.

Answer(s): C



You recently joined a machine learning team that will soon release a new project. As a lead on the project, you are asked to determine the production readiness of the ML components. The team has already tested features and data, model development, and infrastructure. Which additional readiness check should you recommend to the team?

  1. Ensure that training is reproducible.
  2. Ensure that all hyperparameters are tuned.
  3. Ensure that model performance is monitored.
  4. Ensure that feature expectations are captured in the schema.

Answer(s): C



You work for a credit card company and have been asked to create a custom fraud detection model based on historical data using AutoML Tables. You need to prioritize detection of fraudulent transactions while minimizing false positives. Which optimization objective should you use when training the model?

  1. An optimization objective that minimizes Log loss
  2. An optimization objective that maximizes the Precision at a Recall value of 0.50
  3. An optimization objective that maximizes the area under the precision-recall curve (AUC PR) value
  4. An optimization objective that maximizes the area under the receiver operating characteristic curve (AUC ROC) value

Answer(s): C



Page 14 of 69



Post your Comments and Discuss Google Professional Machine Learning Engineer exam with other Community members:

Tina commented on April 09, 2024
Good questions
Anonymous
upvote

Kavah commented on September 29, 2021
Very responsive and cool support team.
UNITED KINGDOM
upvote