Free Google Cloud Certified Professional Data Engineer Exam Braindumps (page: 24)

Page 24 of 68

When running a pipeline that has a BigQuery source, on your local machine, you continue to get permission denied errors.
What could be the reason for that?

  1. Your gcloud does not have access to the BigQuery resources
  2. BigQuery cannot be accessed from local machines
  3. You are missing gcloud on your machine
  4. Pipelines cannot be run locally

Answer(s): A

Explanation:

When reading from a Dataflow source or writing to a Dataflow sink using DirectPipelineRunner, the Cloud Platform account that you configured with the gcloud executable will need access to the corresponding source/sink


Reference:

https://cloud.google.com/dataflow/java- sdk/JavaDoc/com/google/cloud/dataflow/sdk/runners/DirectPipelineRunner



Which of the following is NOT true about Dataflow pipelines?

  1. Dataflow pipelines are tied to Dataflow, and cannot be run on any other runner
  2. Dataflow pipelines can consume data from other Google Cloud services
  3. Dataflow pipelines can be programmed in Java
  4. Dataflow pipelines use a unified programming model, so can work both with streaming and batch data sources

Answer(s): A

Explanation:

Dataflow pipelines can also run on alternate runtimes like Spark and Flink, as they are built using the Apache Beam SDKs


Reference:

https://cloud.google.com/dataflow/



Which action can a Cloud Dataproc Viewer perform?

  1. Submit a job.
  2. Create a cluster.
  3. Delete a cluster.
  4. List the jobs.

Answer(s): D

Explanation:

A Cloud Dataproc Viewer is limited in its actions based on its role. A viewer can only list clusters, get cluster details, list jobs, get job details, list operations, and get operation details.


Reference:

https://cloud.google.com/dataproc/docs/concepts/iam#iam_roles_and_cloud_dataproc_ope rations_summary



How would you query specific partitions in a BigQuery table?

  1. Use the DAY column in the WHERE clause
  2. Use the EXTRACT(DAY) clause
  3. Use the __PARTITIONTIME pseudo-column in the WHERE clause
  4. Use DATE BETWEEN in the WHERE clause

Answer(s): C

Explanation:

Partitioned tables include a pseudo column named _PARTITIONTIME that contains a date- based timestamp for data loaded into the table. To limit a query to particular partitions (such as Jan 1st and 2nd of 2017), use a clause similar to this:
WHERE _PARTITIONTIME BETWEEN TIMESTAMP('2017-01-01') AND TIMESTAMP('2017-01-02')


Reference:

https://cloud.google.com/bigquery/docs/partitioned- tables#the_partitiontime_pseudo_column



Page 24 of 68



Post your Comments and Discuss Google Google Cloud Certified Professional Data Engineer exam with other Community members:

Guru Gee commented on March 23, 2024
Got to buy the full version to get all the answers. But questions are real and I cleared this paper.
INDIA
upvote

Mahitha Govindu commented on February 08, 2024
useful content
Anonymous
upvote

Mahitha Govindu commented on February 08, 2024
Good content
Anonymous
upvote

hello commented on October 31, 2023
good content
Anonymous
upvote

p das commented on December 07, 2023
very good questions
UNITED STATES
upvote

p das commented on December 07, 2023
very good questions
UNITED STATES
upvote

Simon Mukabana commented on November 09, 2023
Did the exam on 3rd and almost all questions came from this dump.
Anonymous
upvote

hello commented on October 31, 2023
good content
Anonymous
upvote

Swati commented on September 11, 2023
Really helpful
Anonymous
upvote