Free Google Cloud Data Engineer Professional Exam Braindumps (page: 6)

Page 6 of 95

Your company is migrating their 30-node Apache Hadoop cluster to the cloud. They want to re-use Hadoop jobs they have already created and minimize the management of the cluster as much as possible. They also want to be able to persist data beyond the life of the cluster.
What should you do?

  1. Create a Google Cloud Dataflow job to process the data.
  2. Create a Google Cloud Dataproc cluster that uses persistent disks for HDFS.
  3. Create a Hadoop cluster on Google Compute Engine that uses persistent disks.
  4. Create a Cloud Dataproc cluster that uses the Google Cloud Storage connector.
  5. Create a Hadoop cluster on Google Compute Engine that uses Local SSD disks.

Answer(s): D



Business owners at your company have given you a database of bank transactions. Each row contains the user ID, transaction type, transaction location, and transaction amount. They ask you to investigate what type of machine learning can be applied to the dat

  1. Which three machine learning applications can you use? (Choose three.)
  2. Supervised learning to determine which transactions are most likely to be fraudulent.
  3. Unsupervised learning to determine which transactions are most likely to be fraudulent.
  4. Clustering to divide the transactions into N categories based on feature similarity.
  5. Supervised learning to predict the location of a transaction.
  6. Reinforcement learning to predict the location of a transaction.
  7. Unsupervised learning to predict the location of a transaction.

Answer(s): B,C,D



Your company's on-premises Apache Hadoop servers are approaching end-of-life, and IT has decided to migrate the cluster to Google Cloud Dataproc. A like-for-like migration of the cluster would require 50 TB of Google Persistent Disk per node. The CIO is concerned about the cost of using that much block storage. You want to minimize the storage cost of the migration.
What should you do?

  1. Put the data into Google Cloud Storage.
  2. Use preemptible virtual machines (VMs) for the Cloud Dataproc cluster.
  3. Tune the Cloud Dataproc cluster so that there is just enough disk for all data.
  4. Migrate some of the cold data into Google Cloud Storage, and keep only the hot data in Persistent Disk.

Answer(s): B


Reference:



You work for a car manufacturer and have set up a data pipeline using Google Cloud Pub/Sub to capture anomalous sensor events. You are using a push subscription in Cloud Pub/Sub that calls a custom HTTPS endpoint that you have created to take action of these anomalous events as they occur. Your custom HTTPS endpoint keeps getting an inordinate amount of duplicate messages.
What is the most likely cause of these duplicate messages?

  1. The message body for the sensor event is too large.
  2. Your custom endpoint has an out-of-date SSL certificate.
  3. The Cloud Pub/Sub topic has too many messages published to it.
  4. Your custom endpoint is not acknowledging messages within the acknowledgement deadline.

Answer(s): B






Post your Comments and Discuss Google Google Cloud Data Engineer Professional exam with other Community members:

Google Cloud Data Engineer Professional Exam Discussions & Posts