Free Associate-Data-Practitioner Exam Braindumps (page: 6)

Page 5 of 19

You need to design a data pipeline that ingests data from CSV, Avro, and Parquet files into Cloud Storage. The data includes raw user input. You need to remove all malicious SQL injections before storing the data in BigQuery.
Which data manipulation methodology should you choose?

  1. EL
  2. ELT
  3. ETL
  4. ETLT

Answer(s): C

Explanation:

The ETL (Extract, Transform, Load) methodology is the best approach for this scenario because it allows you to extract data from the files, transform it by applying the necessary data cleansing (including removing malicious SQL injections), and then load the sanitized data into BigQuery. By transforming the data before loading it into BigQuery, you ensure that only clean and safe data is stored, which is critical for security and data quality.



You are working with a large dataset of customer reviews stored in Cloud Storage. The dataset contains several inconsistencies, such as missing values, incorrect data types, and duplicate entries. You need to clean the data to ensure that it is accurate and consistent before using it for analysis.
What should you do?

  1. Use the PythonOperator in Cloud Composer to clean the data and load it into BigQuery. Use SQL for analysis.
  2. Use BigQuery to batch load the data into BigQuery. Use SQL for cleaning and analysis.
  3. Use Storage Transfer Service to move the data to a different Cloud Storage bucket. Use event triggers to invoke Cloud Run functions to load the data into BigQuery. Use SQL for analysis.
  4. Use Cloud Run functions to clean the data and load it into BigQuery. Use SQL for analysis.

Answer(s): B

Explanation:

Using BigQuery to batch load the data and perform cleaning and analysis with SQL is the best approach for this scenario. BigQuery provides powerful SQL capabilities to handle missing values, enforce correct data types, and remove duplicates efficiently. This method simplifies the pipeline by leveraging BigQuery's built-in processing power for both cleaning and analysis, reducing the need for additional tools or services and minimizing complexity.



Your retail organization stores sensitive application usage data in Cloud Storage. You need to encrypt the data without the operational overhead of managing encryption keys.
What should you do?

  1. Use Google-managed encryption keys (GMEK).
  2. Use customer-managed encryption keys (CMEK).
  3. Use customer-supplied encryption keys (CSEK).
  4. Use customer-supplied encryption keys (CSEK) for the sensitive data and customer-managed encryption keys (CMEK) for the less sensitive data.

Answer(s): A

Explanation:

Using Google-managed encryption keys (GMEK) is the best choice when you want to encrypt sensitive data in Cloud Storage without the operational overhead of managing encryption keys. GMEK is the default encryption mechanism in Google Cloud, and it ensures that data is automatically encrypted at rest with no additional setup or maintenance required. It provides strong security while eliminating the need for manual key management.



You work for a financial organization that stores transaction data in BigQuery. Your organization has a regulatory requirement to retain data for a minimum of seven years for auditing purposes. You need to ensure that the data is retained for seven years using an efficient and cost-optimized approach.
What should you do?

  1. Create a partition by transaction date, and set the partition expiration policy to seven years.
  2. Set the table-level retention policy in BigQuery to seven years.
  3. Set the dataset-level retention policy in BigQuery to seven years.
  4. Export the BigQuery tables to Cloud Storage daily, and enforce a lifecycle management policy that has a seven-year retention rule.

Answer(s): B

Explanation:

Setting a table-level retention policy in BigQuery to seven years is the most efficient and cost- optimized solution to meet the regulatory requirement. A table-level retention policy ensures that the data cannot be deleted or overwritten before the specified retention period expires, providing compliance with auditing requirements while keeping the data within BigQuery for easy access and analysis. This approach avoids the complexity and additional costs of exporting data to Cloud Storage.






Post your Comments and Discuss Google Associate-Data-Practitioner exam with other Community members:

Associate-Data-Practitioner Exam Discussions & Posts