Download AWS Certified Data Engineer - Associate.DEA-C01.ExamTopics.2025-12-11.239q.vcex

Vendor: Amazon
Exam Code: DEA-C01
Exam Name: AWS Certified Data Engineer - Associate
Date: Dec 11, 2025
File Size: 208 KB

How to open VCEX files?

Files with VCEX extension can be opened by ProfExam Simulator.

Demo Questions

Question 1
A financial company wants to implement a data mesh. The data mesh must support centralized data governance, data analysis, and data access control. The company has decided to use AWS Glue for data catalogs and extract, transform, and load (ETL) operations.
Which combination of AWS services will implement a data mesh? (Choose two.)
  1. Use Amazon Aurora for data storage. Use an Amazon Redshift provisioned cluster for data analysis.
  2. Use Amazon S3 for data storage. Use Amazon Athena for data analysis.
  3. Use AWS Glue DataBrew for centralized data governance and access control.
  4. Use Amazon RDS for data storage. Use Amazon EMR for data analysis.
  5. Use AWS Lake Formation for centralized data governance and access control.
Correct answer: BE
Explanation:
BE: 20 - Mosted
BE: 20 - Mosted
Question 2
A company created an extract, transform, and load (ETL) data pipeline in AWS Glue. A data engineer must crawl a table that is in Microsoft SQL Server. The data engineer needs to extract, transform, and load the output of the crawl to an Amazon S3 bucket. The data engineer also must orchestrate the data pipeline.
Which AWS service or feature will meet these requirements MOST cost-effectively?
  1. AWS Step Functions
  2. AWS Glue workflows
  3. AWS Glue Studio
  4. Amazon Managed Workflows for Apache Airflow (Amazon MWAA)
Correct answer: B
Explanation:
B: 21 - Mosted
B: 21 - Mosted
Question 3
A financial services company stores financial data in Amazon Redshift. A data engineer wants to run real-time queries on the financial data to support a web-based trading application. The data engineer wants to run the queries from within the trading application.
Which solution will meet these requirements with the LEAST operational overhead?
  1. Establish WebSocket connections to Amazon Redshift.
  2. Use the Amazon Redshift Data API.
  3. Set up Java Database Connectivity (JDBC) connections to Amazon Redshift.
  4. Store frequently accessed data in Amazon S3. Use Amazon S3 Select to run the queries.
Correct answer: B
Explanation:
B: 17 - Mosted
B: 17 - Mosted
Question 4
A company uses Amazon Athena for one-time queries against data that is in Amazon S3. The company has several use cases. The company must implement permission controls to separate query processes and access to query history among users, teams, and applications that are in the same AWS account.
Which solution will meet these requirements?
  1. Create an S3 bucket for each use case. Create an S3 bucket policy that grants permissions to appropriate individual IAM users. Apply the S3 bucket policy to the S3 bucket.
  2. Create an Athena workgroup for each use case. Apply tags to the workgroup. Create an IAM policy that uses the tags to apply appropriate permissions to the workgroup.
  3. Create an IAM role for each use case. Assign appropriate permissions to the role for each use case. Associate the role with Athena.
  4. Create an AWS Glue Data Catalog resource policy that grants permissions to appropriate individual IAM users for each use case. Apply the resource policy to the specific tables that Athena uses.
Correct answer: B
Explanation:
B: 23 - Mosted
B: 23 - Mosted
Question 5
A data engineer needs to create an AWS Lambda function that converts the format of data from .csv to Apache Parquet. The Lambda function must run only if a user uploads a .csv file to an Amazon S3 bucket.
Which solution will meet these requirements with the LEAST operational overhead?
  1. Create an S3 event notification that has an event type of s3:ObjectCreated:*. Use a filter rule to generate notifications only when the suffix includes .csv. Set the Amazon Resource Name (ARN) of the Lambda function as the destination for the event notification.
  2. Create an S3 event notification that has an event type of s3:ObjectTagging:* for objects that have a tag set to .csv. Set the Amazon Resource Name (ARN) of the Lambda function as the destination for the event notification.
  3. Create an S3 event notification that has an event type of s3:*. Use a filter rule to generate notifications only when the suffix includes .csv. Set the Amazon Resource Name (ARN) of the Lambda function as the destination for the event notification.
  4. Create an S3 event notification that has an event type of s3:ObjectCreated:*. Use a filter rule to generate notifications only when the suffix includes .csv. Set an Amazon Simple Notification Service (Amazon SNS) topic as the destination for the event notification. Subscribe the Lambda function to the SNS topic.
Correct answer: A
Explanation:
A: 19 - Mosted
A: 19 - Mosted
Question 6
A data engineer needs Amazon Athena queries to finish faster. The data engineer notices that all the files the Athena queries use are currently stored in uncompressed .csv format. The data engineer also notices that users perform most queries by selecting a specific column.
Which solution will MOST speed up the Athena query performance?
  1. Change the data format from .csv to JSON format. Apply Snappy compression.
  2. Compress the .csv files by using Snappy compression.
  3. Change the data format from .csv to Apache Parquet. Apply Snappy compression.
  4. Compress the .csv files by using gzip compression.
Correct answer: C
Explanation:
C: 18 - Mosted
C: 18 - Mosted
Question 7
A manufacturing company collects sensor data from its factory floor to monitor and enhance operational efficiency. The company uses Amazon Kinesis Data Streams to publish the data that the sensors collect to a data stream. Then Amazon Kinesis Data Firehose writes the data to an Amazon S3 bucket.
The company needs to display a real-time view of operational efficiency on a large screen in the manufacturing facility.
Which solution will meet these requirements with the LOWEST latency?
  1. Use Amazon Managed Service for Apache Flink (previously known as Amazon Kinesis Data Analytics) to process the sensor data. Use a connector for Apache Flink to write data to an Amazon Timestream database. Use the Timestream database as a source to create a Grafana dashboard.
  2. Configure the S3 bucket to send a notification to an AWS Lambda function when any new object is created. Use the Lambda function to publish the data to Amazon Aurora. Use Aurora as a source to create an Amazon QuickSight dashboard.
  3. Use Amazon Managed Service for Apache Flink (previously known as Amazon Kinesis Data Analytics) to process the sensor data. Create a new Data Firehose delivery stream to publish data directly to an Amazon Timestream database. Use the Timestream database as a source to create an Amazon QuickSight dashboard.
  4. Use AWS Glue bookmarks to read sensor data from the S3 bucket in real time. Publish the data to an Amazon Timestream database. Use the Timestream database as a source to create a Grafana dashboard.
Correct answer: A
Explanation:
A: 34 - MostedC: 1
A: 34 - MostedC: 1
Question 8
A company loads transaction data for each day into Amazon Redshift tables at the end of each day. The company wants to have the ability to track which tables have been loaded and which tables still need to be loaded.
A data engineer wants to store the load statuses of Redshift tables in an Amazon DynamoDB table. The data engineer creates an AWS Lambda function to publish the details of the load statuses to DynamoDB.
How should the data engineer invoke the Lambda function to write load statuses to the DynamoDB table?
  1. Use a second Lambda function to invoke the first Lambda function based on Amazon CloudWatch events.
  2. Use the Amazon Redshift Data API to publish an event to Amazon EventBridge. Configure an EventBridge rule to invoke the Lambda function.
  3. Use the Amazon Redshift Data API to publish a message to an Amazon Simple Queue Service (Amazon SQS) queue. Configure the SQS queue to invoke the Lambda function.
  4. Use a second Lambda function to invoke the first Lambda function based on AWS CloudTrail events.
Correct answer: B
Explanation:
B: 14 - MostedD: 2
B: 14 - MostedD: 2
Question 9
A data engineer needs to securely transfer 5 TB of data from an on-premises data center to an Amazon S3 bucket. Approximately 5% of the data changes every day. Updates to the data need to be regularly proliferated to the S3 bucket. The data includes files that are in multiple formats. The data engineer needs to automate the transfer process and must schedule the process to run periodically.
Which AWS service should the data engineer use to transfer the data in the MOST operationally efficient way?
  1. AWS DataSync
  2. AWS Glue
  3. AWS Direct Connect
  4. Amazon S3 Transfer Acceleration
Correct answer: A
Explanation:
A: 8 - Mosted
A: 8 - Mosted
Question 10
A company uses an on-premises Microsoft SQL Server database to store financial transaction data. The company migrates the transaction data from the on-premises database to AWS at the end of each month. The company has noticed that the cost to migrate data from the on-premises database to an Amazon RDS for SQL Server database has increased recently.
The company requires a cost-effective solution to migrate the data to AWS. The solution must cause minimal downtown for the applications that access the database.
Which AWS service should the company use to meet these requirements?
  1. AWS Lambda
  2. AWS Database Migration Service (AWS DMS)
  3. AWS Direct Connect
  4. AWS DataSync
Correct answer: B
Explanation:
B: 25 - Mosted
B: 25 - Mosted
HOW TO OPEN VCE FILES

Use VCE Exam Simulator to open VCE files
Avanaset

HOW TO OPEN VCEX FILES

Use ProfExam Simulator to open VCEX files
ProfExam Screen

ProfExam
ProfExam at a 20% markdown

You have the opportunity to purchase ProfExam at a 20% reduced price

Get Now!