Free DEA-C01 Exam Braindumps (page: 26)

Page 25 of 53

A lab uses IoT sensors to monitor humidity, temperature, and pressure for a project. The sensors send 100 KB of data every 10 seconds. A downstream process will read the data from an Amazon S3 bucket every 30 seconds.
Which solution will deliver the data to the S3 bucket with the LEAST latency?

  1. Use Amazon Kinesis Data Streams and Amazon Kinesis Data Firehose to deliver the data to the S3 bucket. Use the default buffer interval for Kinesis Data Firehose.
  2. Use Amazon Kinesis Data Streams to deliver the data to the S3 bucket. Configure the stream to use 5 provisioned shards.
  3. Use Amazon Kinesis Data Streams and call the Kinesis Client Library to deliver the data to the S3 bucket. Use a 5 second buffer interval from an application.
  4. Use Amazon Managed Service for Apache Flink (previously known as Amazon Kinesis Data Analytics) and Amazon Kinesis Data Firehose to deliver the data to the S3 bucket. Use a 5 second buffer interval for Kinesis Data Firehose.

Answer(s): C



A company wants to use machine learning (ML) to perform analytics on data that is in an Amazon S3 data lake. The company has two data transformation requirements that will give consumers within the company the ability to create reports.
The company must perform daily transformations on 300 GB of data that is in a variety format that must arrive in Amazon S3 at a scheduled time. The company must perform one-time transformations of terabytes of archived data that is in the S3 data lake. The company uses Amazon Managed Workflows for Apache Airflow (Amazon MWAA) Directed Acyclic Graphs (DAGs) to orchestrate processing.
Which combination of tasks should the company schedule in the Amazon MWAA DAGs to meet these requirements MOST cost-effectively? (Choose two.)

  1. For daily incoming data, use AWS Glue crawlers to scan and identify the schema.
  2. For daily incoming data, use Amazon Athena to scan and identify the schema.
  3. For daily incoming data, use Amazon Redshift to perform transformations.
  4. For daily and archived data, use Amazon EMR to perform data transformations.
  5. For archived data, use Amazon SageMaker to perform data transformations.

Answer(s): A,D



A retail company uses AWS Glue for extract, transform, and load (ETL) operations on a dataset that contains information about customer orders. The company wants to implement specific validation rules to ensure data accuracy and consistency.
Which solution will meet these requirements?

  1. Use AWS Glue job bookmarks to track the data for accuracy and consistency.
  2. Create custom AWS Glue Data Quality rulesets to define specific data quality checks.
  3. Use the built-in AWS Glue Data Quality transforms for standard data quality validations.
  4. Use AWS Glue Data Catalog to maintain a centralized data schema and metadata repository.

Answer(s): B



An insurance company stores transaction data that the company compressed with gzip.
The company needs to query the transaction data for occasional audits.
Which solution will meet this requirement in the MOST cost-effective way?

  1. Store the data in Amazon Glacier Flexible Retrieval. Use Amazon S3 Glacier Select to query the data.
  2. Store the data in Amazon S3. Use Amazon S3 Select to query the data.
  3. Store the data in Amazon S3. Use Amazon Athena to query the data.
  4. Store the data in Amazon Glacier Instant Retrieval. Use Amazon Athena to query the data.

Answer(s): B






Post your Comments and Discuss Amazon DEA-C01 exam with other Community members:

DEA-C01 Discussions & Posts