Free Google Cloud Data Engineer Professional Exam Braindumps (page: 5)

Page 5 of 95

You want to process payment transactions in a point-of-sale application that will run on Google Cloud Platform. Your user base could grow exponentially, but you do not want to manage infrastructure scaling.

Which Google database service should you use?

  1. Cloud SQL
  2. BigQuery
  3. Cloud Bigtable
  4. Cloud Datastore

Answer(s): A



You want to use a database of information about tissue samples to classify future tissue samples as either normal or mutated. You are evaluating an unsupervised anomaly detection method for classifying the tissue samples.
Which two characteristic support this method? (Choose two.)

  1. There are very few occurrences of mutations relative to normal samples.
  2. There are roughly equal occurrences of both normal and mutated samples in the database.
  3. You expect future mutations to have different features from the mutated samples in the database.
  4. You expect future mutations to have similar features to the mutated samples in the database.
  5. You already have labels for which samples are mutated and which are normal in the database.

Answer(s): A,D

Explanation:

Unsupervised anomaly detection techniques detect anomalies in an unlabeled test data set under the assumption that the majority of the instances in the data set are normal by looking for instances that seem to fit least to the remainder of the data set.
https://en.wikipedia.org/wiki/Anomaly_detection



You need to store and analyze social media postings in Google BigQuery at a rate of 10,000 messages per minute in near real-time. Initially, design the application to use streaming inserts for individual postings. Your application also performs data aggregations right after the streaming inserts. You discover that the queries after streaming inserts do not exhibit strong consistency, and reports from the queries might miss in-flight dat

  1. How can you adjust your application design?
  2. Re-write the application to load accumulated data every 2 minutes.
  3. Convert the streaming insert code to batch load for individual messages.
  4. Load the original message to Google Cloud SQL, and export the table every hour to BigQuery via streaming inserts.
  5. Estimate the average latency for data availability after streaming inserts, and always run queries after waiting twice as long.

Answer(s): D

Explanation:

The data is first comes to buffer and then written to Storage. If we are running queries in buffer we will face above mentioned issues. If we wait for the bigquery to write the data to storage then we won't face the issue. So We need to wait till it's written tio storage



Your startup has never implemented a formal security policy. Currently, everyone in the company has access to the datasets stored in Google BigQuery. Teams have freedom to use the service as they see fit, and they have not documented their use cases. You have been asked to secure the data warehouse. You need to discover what everyone is doing.
What should you do first?

  1. Use Google Stackdriver Audit Logs to review data access.
  2. Get the identity and access management IIAM) policy of each table
  3. Use Stackdriver Monitoring to see the usage of BigQuery query slots.
  4. Use the Google Cloud Billing API to see what account the warehouse is being billed to.

Answer(s): A






Post your Comments and Discuss Google Google Cloud Data Engineer Professional exam with other Community members:

Google Cloud Data Engineer Professional Exam Discussions & Posts