Free Professional Data Engineer Exam Braindumps (page: 24)

Page 24 of 68

When running a pipeline that has a BigQuery source, on your local machine, you continue to get permission denied errors.
What could be the reason for that?

  1. Your gcloud does not have access to the BigQuery resources
  2. BigQuery cannot be accessed from local machines
  3. You are missing gcloud on your machine
  4. Pipelines cannot be run locally

Answer(s): A

Explanation:

When reading from a Dataflow source or writing to a Dataflow sink using DirectPipelineRunner, the Cloud Platform account that you configured with the gcloud executable will need access to the corresponding source/sink


Reference:

https://cloud.google.com/dataflow/java- sdk/JavaDoc/com/google/cloud/dataflow/sdk/runners/DirectPipelineRunner



Which of the following is NOT true about Dataflow pipelines?

  1. Dataflow pipelines are tied to Dataflow, and cannot be run on any other runner
  2. Dataflow pipelines can consume data from other Google Cloud services
  3. Dataflow pipelines can be programmed in Java
  4. Dataflow pipelines use a unified programming model, so can work both with streaming and batch data sources

Answer(s): A

Explanation:

Dataflow pipelines can also run on alternate runtimes like Spark and Flink, as they are built using the Apache Beam SDKs


Reference:

https://cloud.google.com/dataflow/



Which action can a Cloud Dataproc Viewer perform?

  1. Submit a job.
  2. Create a cluster.
  3. Delete a cluster.
  4. List the jobs.

Answer(s): D

Explanation:

A Cloud Dataproc Viewer is limited in its actions based on its role. A viewer can only list clusters, get cluster details, list jobs, get job details, list operations, and get operation details.


Reference:

https://cloud.google.com/dataproc/docs/concepts/iam#iam_roles_and_cloud_dataproc_ope rations_summary



How would you query specific partitions in a BigQuery table?

  1. Use the DAY column in the WHERE clause
  2. Use the EXTRACT(DAY) clause
  3. Use the __PARTITIONTIME pseudo-column in the WHERE clause
  4. Use DATE BETWEEN in the WHERE clause

Answer(s): C

Explanation:

Partitioned tables include a pseudo column named _PARTITIONTIME that contains a date- based timestamp for data loaded into the table. To limit a query to particular partitions (such as Jan 1st and 2nd of 2017), use a clause similar to this:
WHERE _PARTITIONTIME BETWEEN TIMESTAMP('2017-01-01') AND TIMESTAMP('2017-01-02')


Reference:

https://cloud.google.com/bigquery/docs/partitioned- tables#the_partitiontime_pseudo_column



Page 24 of 68



Post your Comments and Discuss Google Professional Data Engineer exam with other Community members:

madhan commented on June 16, 2023
next question
EUROPEAN UNION
upvote