Amazon AWS Certified Data Engineer - Associate DEA-C01 Exam
AWS Certified Data Engineer - Associate DEA-C01 (Page 6 )

Updated On: 1-Feb-2026

A data engineer needs to use Amazon Neptune to develop graph applications.
Which programming languages should the engineer use to develop the graph applications? (Choose two.)

  1. Gremlin
  2. SQL
  3. ANSI SQL
  4. SPARQL
  5. Spark SQL

Answer(s): A,D



A mobile gaming company wants to capture data from its gaming app. The company wants to make the data available to three internal consumers of the data. The data records are approximately 20 KB in size.
The company wants to achieve optimal throughput from each device that runs the gaming app. Additionally, the company wants to develop an application to process data streams. The stream-processing application must have dedicated throughput for each internal consumer.
Which solution will meet these requirements?

  1. Configure the mobile app to call the PutRecords API operation to send data to Amazon Kinesis Data Streams. Use the enhanced fan-out feature with a stream for each internal consumer.
  2. Configure the mobile app to call the PutRecordBatch API operation to send data to Amazon Kinesis Data Firehose. Submit an AWS Support case to turn on dedicated throughput for the company’s AWS account. Allow each internal consumer to access the stream.
  3. Configure the mobile app to use the Amazon Kinesis Producer Library (KPL) to send data to Amazon Kinesis Data Firehose. Use the enhanced fan-out feature with a stream for each internal consumer.
  4. Configure the mobile app to call the PutRecords API operation to send data to Amazon Kinesis Data Streams. Host the stream-processing application for each internal consumer on Amazon EC2 instances. Configure auto scaling for the EC2 instances.

Answer(s): A



A retail company uses an Amazon Redshift data warehouse and an Amazon S3 bucket. The company ingests retail order data into the S3 bucket every day.
The company stores all order data at a single path within the S3 bucket. The data has more than 100 columns. The company ingests the order data from a third-party application that generates more than 30 files in CSV format every day. Each CSV file is between 50 and 70 MB in size.
The company uses Amazon Redshift Spectrum to run queries that select sets of columns. Users aggregate metrics based on daily orders. Recently, users have reported that the performance of the queries has degraded. A data engineer must resolve the performance issues for the queries.
Which combination of steps will meet this requirement with LEAST developmental effort? (Choose two.)

  1. Configure the third-party application to create the files in a columnar format.
  2. Develop an AWS Glue ETL job to convert the multiple daily CSV files to one file for each day.
  3. Partition the order data in the S3 bucket based on order date.
  4. Configure the third-party application to create the files in JSON format.
  5. Load the JSON data into the Amazon Redshift table in a SUPER type column.

Answer(s): A,C



Files from multiple data sources arrive in an Amazon S3 bucket on a regular basis. A data engineer wants to ingest new files into Amazon Redshift in near real time when the new files arrive in the S3 bucket.
Which solution will meet these requirements?

  1. Use the query editor v2 to schedule a COPY command to load new files into Amazon Redshift.
  2. Use the zero-ETL integration between Amazon Aurora and Amazon Redshift to load new files into Amazon Redshift.
  3. Use AWS Glue job bookmarks to extract, transform, and load (ETL) load new files into Amazon Redshift.
  4. Use S3 Event Notifications to invoke an AWS Lambda function that loads new files into Amazon Redshift.

Answer(s): D



A company stores customer records in Amazon S3. The company must not delete or modify the customer record data for 7 years after each record is created. The root user also must not have the ability to delete or modify the data.
A data engineer wants to use S3 Object Lock to secure the data.
Which solution will meet these requirements?

  1. Enable governance mode on the S3 bucket. Use a default retention period of 7 years.
  2. Enable compliance mode on the S3 bucket. Use a default retention period of 7 years.
  3. Place a legal hold on individual objects in the S3 bucket. Set the retention period to 7 years.
  4. Set the retention period for individual objects in the S3 bucket to 7 years.

Answer(s): B



Viewing page 6 of 43
Viewing questions 26 - 30 out of 298 questions



Post your Comments and Discuss Amazon AWS Certified Data Engineer - Associate DEA-C01 exam prep with other Community members:

Join the AWS Certified Data Engineer - Associate DEA-C01 Discussion