Free DAS-C01 Exam Braindumps (page: 14)

Page 13 of 42

A large company has a central data lake to run analytics across different departments. Each department uses a separate AWS account and stores its data in an
Amazon S3 bucket in that account. Each AWS account uses the AWS Glue Data Catalog as its data catalog. There are different data lake access requirements based on roles. Associate analysts should only have read access to their departmental data. Senior data analysts can have access in multiple departments including theirs, but for a subset of columns only. Which solution achieves these required access patterns to minimize costs and administrative tasks?

  1. Consolidate all AWS accounts into one account. Create different S3 buckets for each department and move all the data from every account to the central data lake account. Migrate the individual data catalogs into a central data catalog and apply ne-grained permissions to give to each user the required access to tables and databases in AWS Glue and Amazon S3.
  2. Keep the account structure and the individual AWS Glue catalogs on each account. Add a central data lake account and use AWS Glue to catalog data from various accounts. Con gure cross-account access for AWS Glue crawlers to scan the data in each departmental S3 bucket to identify the schema and populate the catalog. Add the senior data analysts into the central account and apply highly detailed access controls in the Data Catalog and Amazon S3.
  3. Set up an individual AWS account for the central data lake. Use AWS Lake Formation to catalog the cross-account locations. On each individual S3 bucket, modify the bucket policy to grant S3 permissions to the Lake Formation service-linked role. Use Lake Formation permissions to add ne-grained access controls to allow senior analysts to view speci c tables and columns.
  4. Set up an individual AWS account for the central data lake and con gure a central S3 bucket. Use an AWS Lake Formation blueprint to move the data from the various buckets into the central S3 bucket. On each individual bucket, modify the bucket policy to grant S3 permissions to the Lake Formation service-linked role. Use Lake Formation permissions to add ne-grained access controls for both associate and senior analysts to view speci c tables and columns.

Answer(s): C



A company wants to improve user satisfaction for its smart home system by adding more features to its recommendation engine. Each sensor asynchronously pushes its nested JSON data into Amazon Kinesis Data Streams using the Kinesis Producer Library (KPL) in Java. Statistics from a set of failed sensors showed that, when a sensor is malfunctioning, its recorded data is not always sent to the cloud. The company needs a solution that offers near-real-time analytics on the data from the most updated sensors.
Which solution enables the company to meet these requirements?

  1. Set the RecordMaxBufferedTime property of the KPL to "1'^ " to disable the buffering on the sensor side. Use Kinesis Data Analytics to enrich the data based on a company-developed anomaly detection SQL script. Push the enriched data to a eet of Kinesis data streams and enable the data transformation feature to atten the JSON le. Instantiate a dense storage Amazon Redshift cluster and use it as the destination for the Kinesis Data Firehose delivery stream.
  2. Update the sensors code to use the PutRecord/PutRecords call from the Kinesis Data Streams API with the AWS SDK for Java. Use Kinesis Data Analytics to enrich the data based on a company-developed anomaly detection SQL script. Direct the output of KDA application to a Kinesis Data Firehose delivery stream, enable the data transformation feature to atten the JSON le, and set the Kinesis Data Firehose destination to an Amazon OpenSearch Service (Amazon Elasticsearch Service) cluster.
  3. Set the RecordMaxBufferedTime property of the KPL to "0" to disable the buffering on the sensor side. Connect for each stream a dedicated Kinesis Data Firehose delivery stream and enable the data transformation feature to atten the JSON le before sending it to an Amazon S3 bucket. Load the S3 data into an Amazon Redshift cluster.
  4. Update the sensors code to use the PutRecord/PutRecords call from the Kinesis Data Streams API with the AWS SDK for Java. Use AWS Glue to fetch and process data from the stream using the Kinesis Client Library (KCL). Instantiate an Amazon Elasticsearch Service cluster and use AWS Lambda to directly push data into it.

Answer(s): B



A global company has different sub-organizations, and each sub-organization sells its products and services in various countries. The company's senior leadership wants to quickly identify which sub-organization is the strongest performer in each country. All sales data is stored in Amazon S3 in Parquet format.
Which approach can provide the visuals that senior leadership requested with the least amount of effort?

  1. Use Amazon QuickSight with Amazon Athena as the data source. Use heat maps as the visual type.
  2. Use Amazon QuickSight with Amazon S3 as the data source. Use heat maps as the visual type.
  3. Use Amazon QuickSight with Amazon Athena as the data source. Use pivot tables as the visual type.
  4. Use Amazon QuickSight with Amazon S3 as the data source. Use pivot tables as the visual type.

Answer(s): A



A company has 1 million scanned documents stored as image les in Amazon S3. The documents contain typewritten application forms with information including the applicant rst name, applicant last name, application date, application type, and application text. The company has developed a machine learning algorithm to extract the metadata values from the scanned documents. The company wants to allow internal data analysts to analyze and nd applications using the applicant name, application date, or application text. The original images should also be downloadable. Cost control is secondary to query performance.
Which solution organizes the images and metadata to drive insights while meeting the requirements?

  1. For each image, use object tags to add the metadata. Use Amazon S3 Select to retrieve the les based on the applicant name and application date.
  2. Index the metadata and the Amazon S3 location of the image le in Amazon OpenSearch Service (Amazon Elasticsearch Service). Allow the data analysts to use OpenSearch Dashboards (Kibana) to submit queries to the Amazon OpenSearch Service (Amazon Elasticsearch Service) cluster.
  3. Store the metadata and the Amazon S3 location of the image le in an Amazon Redshift table. Allow the data analysts to run ad-hoc queries on the table.
  4. Store the metadata and the Amazon S3 location of the image les in an Apache Parquet le in Amazon S3, and de ne a table in the AWS Glue Data Catalog. Allow data analysts to use Amazon Athena to submit custom queries.

Answer(s): B






Post your Comments and Discuss Amazon DAS-C01 exam with other Community members:

DAS-C01 Discussions & Posts