Free DAS-C01 Exam Braindumps (page: 6)

Page 5 of 42

A team of data scientists plans to analyze market trend data for their company's new investment strategy. The trend data comes from ve different data sources in large volumes. The team wants to utilize Amazon Kinesis to support their use case. The team uses SQL-like queries to analyze trends and wants to send noti cations based on certain signi cant patterns in the trends. Additionally, the data scientists want to save the data to Amazon S3 for archival and historical re- processing, and use AWS managed services wherever possible. The team wants to implement the lowest-cost solution.
Which solution meets these requirements?

  1. Publish data to one Kinesis data stream. Deploy a custom application using the Kinesis Client Library (KCL) for analyzing trends, and send noti cations using Amazon SNS. Con gure Kinesis Data Firehose on the Kinesis data stream to persist data to an S3 bucket.
  2. Publish data to one Kinesis data stream. Deploy Kinesis Data Analytic to the stream for analyzing trends, and con gure an AWS Lambda function as an output to send noti cations using Amazon SNS. Con gure Kinesis Data Firehose on the Kinesis data stream to persist data to an S3 bucket.
  3. Publish data to two Kinesis data streams. Deploy Kinesis Data Analytics to the rst stream for analyzing trends, and con gure an AWS Lambda function as an output to send noti cations using Amazon SNS. Con gure Kinesis Data Firehose on the second Kinesis data stream to persist data to an S3 bucket.
  4. Publish data to two Kinesis data streams. Deploy a custom application using the Kinesis Client Library (KCL) to the rst stream for analyzing trends, and send noti cations using Amazon SNS. Con gure Kinesis Data Firehose on the second Kinesis data stream to persist data to an S3 bucket.

Answer(s): A



A company currently uses Amazon Athena to query its global datasets. The regional data is stored in Amazon S3 in the us-east-1 and us-west-2 Regions. The data is not encrypted. To simplify the query process and manage it centrally, the company wants to use Athena in us-west-2 to query data from Amazon S3 in both
Regions. The solution should be as low-cost as possible.
What should the company do to achieve this goal?

  1. Use AWS DMS to migrate the AWS Glue Data Catalog from us-east-1 to us-west-2. Run Athena queries in us-west-2.
  2. Run the AWS Glue crawler in us-west-2 to catalog datasets in all Regions. Once the data is crawled, run Athena queries in us-west-2.
  3. Enable cross-Region replication for the S3 buckets in us-east-1 to replicate data in us-west-2. Once the data is replicated in us-west-2, run the AWS Glue crawler there to update the AWS Glue Data Catalog in us-west-2 and run Athena queries.
  4. Update AWS Glue resource policies to provide us-east-1 AWS Glue Data Catalog access to us-west-2. Once the catalog in us-west-2 has access to the catalog in us-east-1, run Athena queries in us-west-2.

Answer(s): B



A large company receives les from external parties in Amazon EC2 throughout the day. At the end of the day, the les are combined into a single le, compressed into a gzip le, and uploaded to Amazon S3. The total size of all the les is close to 100 GB daily. Once the les are uploaded to Amazon S3, an
AWS Batch program executes a COPY command to load the les into an Amazon Redshift cluster.
Which program modi cation will accelerate the COPY process?

  1. Upload the individual les to Amazon S3 and run the COPY command as soon as the les become available.
  2. Split the number of les so they are equal to a multiple of the number of slices in the Amazon Redshift cluster. Gzip and upload the les to Amazon S3. Run the COPY command on the les.
  3. Split the number of les so they are equal to a multiple of the number of compute nodes in the Amazon Redshift cluster. Gzip and upload the les to Amazon S3. Run the COPY command on the les.
  4. Apply sharding by breaking up the les so the distkey columns with the same values go to the same le. Gzip and upload the sharded les to Amazon S3. Run the COPY command on the les.

Answer(s): B


Reference:

https://docs.aws.amazon.com/redshift/latest/dg/t_splitting-data- les.html



A large ride-sharing company has thousands of drivers globally serving millions of unique customers every day. The company has decided to migrate an existing data mart to Amazon Redshift. The existing schema includes the following tables.
A trips fact table for information on completed rides.
A drivers dimension table for driver pro les.
A customers fact table holding customer pro le information.
The company analyzes trip details by date and destination to examine pro tability by region. The drivers data rarely changes. The customers data frequently changes.
What table design provides optimal query performance?

  1. Use DISTSTYLE KEY (destination) for the trips table and sort by date. Use DISTSTYLE ALL for the drivers and customers tables.
  2. Use DISTSTYLE EVEN for the trips table and sort by date. Use DISTSTYLE ALL for the drivers table. Use DISTSTYLE EVEN for the customers table.
  3. Use DISTSTYLE KEY (destination) for the trips table and sort by date. Use DISTSTYLE ALL for the drivers table. Use DISTSTYLE EVEN for the customers table.
  4. Use DISTSTYLE EVEN for the drivers table and sort by date. Use DISTSTYLE ALL for both fact tables.

Answer(s): C






Post your Comments and Discuss Amazon DAS-C01 exam with other Community members:

DAS-C01 Discussions & Posts