Free Professional Data Engineer Exam Braindumps (page: 24)

Page 23 of 95

Which Java SDK class can you use to run your Dataflow programs locally?

  1. LocalRunner
  2. DirectPipelineRunner
  3. MachineRunner
  4. LocalPipelineRunner

Answer(s): B

Explanation:

DirectPipelineRunner allows you to execute operations in the pipeline directly, without any optimization. Useful for small local execution and tests


Reference:

https://cloud.google.com/dataflow/java-
sdk/JavaDoc/com/google/cloud/dataflow/sdk/runners/DirectPipelineRunner



The Dataflow SDKs have been recently transitioned into which Apache service?

  1. Apache Spark
  2. Apache Hadoop
  3. Apache Kafka
  4. Apache Beam

Answer(s): D

Explanation:

Dataflow SDKs are being transitioned to Apache Beam, as per the latest Google directive


Reference:

https://cloud.google.com/dataflow/docs/



The _________ for Cloud Bigtable makes it possible to use Cloud Bigtable in a Cloud Dataflow pipeline.

  1. Cloud Dataflow connector
  2. DataFlow SDK
  3. BiqQuery API
  4. BigQuery Data Transfer Service

Answer(s): A

Explanation:

The Cloud Dataflow connector for Cloud Bigtable makes it possible to use Cloud Bigtable in a Cloud Dataflow pipeline. You can use the connector for both batch and streaming operations.


Reference:

https://cloud.google.com/bigtable/docs/dataflow-hbase



Does Dataflow process batch data pipelines or streaming data pipelines?

  1. Only Batch Data Pipelines
  2. Both Batch and Streaming Data Pipelines
  3. Only Streaming Data Pipelines
  4. None of the above

Answer(s): B

Explanation:

Dataflow is a unified processing model, and can execute both streaming and batch data pipelines


Reference:

https://cloud.google.com/dataflow/






Post your Comments and Discuss Google Professional Data Engineer exam with other Community members:

Exam Discussions & Posts