






Study with the several resources on Docsity
Earn points by helping other students or get them with a premium plan
Prepare for your exams
Study with the several resources on Docsity
Earn points to download
Earn points by helping other students or get them with a premium plan
Community
Ask the community for help and clear up your study doubts
Discover the best universities in your country according to Docsity users
Free resources
Download our free guides on studying techniques, anxiety management strategies, and thesis advice from Docsity tutors
A comprehensive set of 390+ exam questions to help candidates prepare for the aws data engineer associate (dea-c01) certification exam. The questions cover a wide range of topics related to aws data engineering, including kinesis data streams, sqs, glue, redshift, and more. Detailed explanations for each question, which can help candidates better understand the concepts and improve their chances of passing the exam on the first attempt. Skillcertpro, the provider of these exam questions, offers a 100% pass guarantee and updates the questions every 2 weeks to ensure that the content remains up-to-date. The document also includes information on the recommended score to achieve before attempting the real exam, as well as the benefits of lifetime access and free updates.
Typology: Exercises
1 / 10
This page cannot be seen from the preview
Don't miss anything!
AWS Data Engineer Associate (DEA-C01) Practice Tests 2024. Contains 390+ exam questions to pass the exam in first attempt. SkillCertPro offers real exam questions for practice for all major IT certifications.
For a full set of 39 0 + questions. Go to https://skillcertpro.com/product/aws-data-engineer-associate-dea-c01-exam- questions/ SkillCertPro offers detailed explanations to each question which helps to understand the concepts better. It is recommended to score above 85% in SkillCertPro exams before attempting a real exam. SkillCertPro updates exam questions every 2 weeks. You will get life time access and life time free updates SkillCertPro assures 100% pass guarantee in first attempt.
Below are the free 10 sample questions.
A data engineer working for an analytics company is working on a consumer to a Kinesis Data Streams application. They have written the consumer using Kinesis Client Library (KCL), however, currently they are receiving an ExpiredIteratorException when reading records from Kinesis Data Streams. What would you recommend to the engineer to solve their issue?
A. Change the capacity mode of the Kinesis Data Stream to on-demand.
B. Increase WCU in DynamoDB checkpointing table.
C. Increase the amount of shards in Kinesis Data Streams.
D. Increase RCU in DynamoDB checkpointing table.
A) Increase WCU in DynamoDB checkpointing table: Kinesis Client Library leverages DynamoDB for coordination and checkpointing. A new shard iterator is returned by every GetRecords request (as NextShardIterator), which you then use in the next GetRecords request (as ShardIterator). If the shard iterator expires immediately, before you can use it, this might indicate that the DynamoDB table used by Kinesis does not have enough capacity.
For more on troubleshooting AWS Kinesis Data Streams: https://docs.aws.amazon.com/streams/latest/dev/troubleshooting- consumers.html#shard-iterator-expires-unexpectedly
You are building a pipeline to process, analyze and classify images. Your image datasets contain images that you need to preprocess as a first step by resizing and enhancing image contrast. Which AWS service should you consider to use to preprocess the datasets?
A. AWS Step Functions
B. AWS Data Pipeline
C. AWS SageMaker Data Wrangler
D. AWS Glue DataBrew
https://docs.aws.amazon.com/databrew/latest/dg/recipe-actions-reference.html https://aws.amazon.com/blogs/machine-learning/prepare-image-data-with- amazon-sagemaker-data-wrangler/
https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGui de/sqs-visibility-timeout.html
Which of the following services are capable of reading from AWS Kinesis Data Streams (SELECT THREE)?
A. Amazon Managed Service for Apache Flink
B. EFS
C. S
D. EC
E. EMR
The currently available services for processing data from Kinesis Data Streams are: Amazon Managed Service for Apache Flink, Spark on Amazon EMR, EC2, Lambda, Kinesis Data Firehose, and the Kinesis Client Library.
S3 and EFS can be output locations using the previously mentioned services, however, they cannot be used with Kinesis Data Streams without an intermediary processing step/service.
For more on AWS Kinesis Data Streams:
https://aws.amazon.com/kinesis/data-streams/
For a full set of 390+ questions. Go to https://skillcertpro.com/product/aws-data-engineer-associate-dea-c01-exam- questions/ SkillCertPro offers detailed explanations to each question which helps to understand the concepts better. It is recommended to score above 85% in SkillCertPro exams before attempting a real exam. SkillCertPro updates exam questions every 2 weeks. You will get life time access and life time free updates SkillCertPro assures 100% pass guarantee in first attempt.
A company has a daily running ETL process, which processes transactions from a production database. The process is not time sensitive and can be run at any point of the day. Currently, the company is in the process of migrating the ETL job to an AWS Glue Spark job. As a Certified Data Engineer, what would be the most cost-efficient way to structure the Glue ETL job?
A. Set the Glue to version 2.
B. Set the execution class of the Glue job to FLEX
C. Set the execution class of the Glue job to STANDARD
D. Set the Glue job to use Spot instances
Which of the following statements are CORRECT regarding AWS SQS (SELECT TWO)?
A. Message size is limited to 256KB
B. Messages cannot be duplicated
B. AWS Lambda C. AWS Data Pipeline D. AWS Step Functions
AWS Glue : Explanation: AWS Glue is the best choice as it provides a comprehensive solution for data integration and transformation. It offers data cataloging, ETL job orchestration, and serverless execution capabilities, making it ideal for efficient, automated, and scalable workflows in a data lake. See: https://docs.aws.amazon.com/glue/latest/dg/what-is-glue.html
AWS Data Pipeline. Explanation: While AWS Data Pipeline can orchestrate data workflows, it‘s not as feature-rich as AWS Glue when it comes to ETL and serverless execution, making it less suitable for the described scenario.
AWS Step Functions. Explanation: AWS Step Functions are great for orchestrating serverless workflows, but do not offer the full range of ETL and data cataloging capabilities required for a data lake integration and transformation solution.
AWS Lambda: Explanation. AWS Lambda is a serverless compute service, but is not specifically designed for data integration and ETL tasks. It can be part of a larger solution, but would require more custom development compared to AWS Glue.
A data engineer working in an analytics company has been tasked to migrate their Apache Cassandra database to AWS. What AWS service should the engineer use to migrate the database to AWS, with the LEAST amount of operational overhead?
A. DocumentDB
B. Amazon Neptune C. Amazon Keyspaces D. Amazon RDS
A data analyst at a social media company wants to create a new Redshift table from a query. What would you recommend to the analyst?
A. Use the SELECT INTO command on Redshift to query data and create a table from the results of the query
B. Use the CREATE TABLE command on Redshift to create a table from a given query
C. Use the COPY command on Redshift to create a table from a given query
D. Use the SELECT command on Redshift and save the intermediary results to S3. E. Use the COPY command to create a new table on Redshift from the S3 data
Use the SELECT INTO command on Redshift to query data and create a table from the results of the query : The SELECT INTO command selects rows defined by any query and inserts them into a new table. You can specify whether to create a temporary or a persistent table. For more on Redshift SELECT INTO: https://docs.aws.amazon.com/redshift/latest/dg/r_SELECT_INTO.ht