Free Professional Data Engineer Exam Braindumps (page: 33)

Page 33 of 68

What is the general recommendation when designing your row keys for a Cloud Bigtable schema?

  1. Include multiple time series values within the row key
  2. Keep the row keep as an 8 bit integer
  3. Keep your row key reasonably short
  4. Keep your row key as long as the field permits

Answer(s): C

Explanation:

A general guide is to, keep your row keys reasonably short. Long row keys take up additional memory and storage and increase the time it takes to get responses from the Cloud Bigtable server.


Reference:

https://cloud.google.com/bigtable/docs/schema-design#row-keys



Which of the following statements is NOT true regarding Bigtable access roles?

  1. Using IAM roles, you cannot give a user access to only one table in a project, rather than all tables in a project.
  2. To give a user access to only one table in a project, grant the user the Bigtable Editor role for
    that table.
  3. You can configure access control only at the project level.
  4. To give a user access to only one table in a project, you must configure access through your application.

Answer(s): B

Explanation:

For Cloud Bigtable, you can configure access control at the project level. For example, you can grant the ability to:
Read from, but not write to, any table within the project. Read from and write to any table within the project, but not manage instances. Read from and write to any table within the project, and manage instances.


Reference:

https://cloud.google.com/bigtable/docs/access-control



If you're running a performance test that depends upon Cloud Bigtable, all the choices except one below are recommended steps.
Which is NOT a recommended step to follow?

  1. Do not use a production instance.
  2. Run your test for at least 10 minutes.
  3. Before you test, run a heavy pre-test for several minutes.
  4. Use at least 300 GB of data.

Answer(s): A

Explanation:

If you're running a performance test that depends upon Cloud Bigtable, be sure to follow these steps as you plan and execute your test:
Use a production instance. A development instance will not give you an accurate sense of how a production instance performs under load.
Use at least 300 GB of data. Cloud Bigtable performs best with 1 TB or more of data. However, 300 GB of data is enough to provide reasonable results in a performance test on
a 3-node cluster. On larger clusters, use 100 GB of data per node. Before you test, run a heavy pre-test for several minutes. This step gives Cloud Bigtable a chance to balance data across your nodes based on the access patterns it observes. Run your test for at least 10 minutes. This step lets Cloud Bigtable further optimize your data, and it helps ensure that you will test reads from disk as well as cached reads from memory.


Reference:

https://cloud.google.com/bigtable/docs/performance



You operate a logistics company, and you want to improve event delivery reliability for vehicle-based sensors. You operate small data centers around the world to capture these events, but leased lines that provide connectivity from your event collection infrastructure to your event processing infrastructure are unreliable, with unpredictable latency. You want to address this issue in the most cost-effective way.
What should you do?

  1. Deploy small Kafka clusters in your data centers to buffer events.
  2. Have the data acquisition devices publish data to Cloud Pub/Sub.
  3. Establish a Cloud Interconnect between all remote data centers and Google.
  4. Write a Cloud Dataflow pipeline that aggregates all data in session windows.

Answer(s): B



Page 33 of 68



Post your Comments and Discuss Google Professional Data Engineer exam with other Community members:

madhan commented on June 16, 2023
next question
EUROPEAN UNION
upvote