Free Professional Data Engineer Exam Braindumps (page: 23)

Page 22 of 95

You have a job that you want to cancel. It is a streaming pipeline, and you want to ensure that any data that is in-flight is processed and written to the output.
Which of the following commands can you use on the Dataflow monitoring console to stop the pipeline job?

  1. Cancel
  2. Drain
  3. Stop
  4. Finish

Answer(s): B

Explanation:

Using the Drain option to stop your job tells the Dataflow service to finish your job in its current state. Your job will immediately stop ingesting new data from input sources, but the Dataflow service will preserve any existing resources (such as worker instances) to finish processing and writing any buffered data in your pipeline.


Reference:

https://cloud.google.com/dataflow/pipelines/stopping-a-pipeline



When running a pipeline that has a BigQuery source, on your local machine, you continue to get permission denied errors.
What could be the reason for that?

  1. Your gcloud does not have access to the BigQuery resources
  2. BigQuery cannot be accessed from local machines
  3. You are missing gcloud on your machine
  4. Pipelines cannot be run locally

Answer(s): A

Explanation:

When reading from a Dataflow source or writing to a Dataflow sink using DirectPipelineRunner, the Cloud Platform account that you configured with the gcloud executable will need access to the corresponding source/sink


Reference:

https://cloud.google.com/dataflow/java-
sdk/JavaDoc/com/google/cloud/dataflow/sdk/runners/DirectPipelineRunner



What Dataflow concept determines when a Window's contents should be output based on certain criteria being met?

  1. Sessions
  2. OutputCriteria
  3. Windows
  4. Triggers

Answer(s): D

Explanation:

Triggers control when the elements for a specific key and window are output. As elements arrive, they are put into one or more windows by a Window transform and its associated WindowFn, and then passed to the associated Trigger to determine if the Windows contents should be output.


Reference:

https://cloud.google.com/dataflow/java-
sdk/JavaDoc/com/google/cloud/dataflow/sdk/transforms/windowing/Trigger



Which of the following is NOT one of the three main types of triggers that Dataflow supports?

  1. Trigger based on element size in bytes
  2. Trigger that is a combination of other triggers
  3. Trigger based on element count
  4. Trigger based on time

Answer(s): A

Explanation:

There are three major kinds of triggers that Dataflow supports: 1. Time-based triggers 2. Data-driven triggers. You can set a trigger to emit results from a window when that window has received a certain number of data elements. 3. Composite triggers. These triggers combine multiple time-based or data-driven triggers in some logical way


Reference:

https://cloud.google.com/dataflow/model/triggers






Post your Comments and Discuss Google Professional Data Engineer exam with other Community members:

Exam Discussions & Posts