AWS β’ DEA-C01
Validates ability to implement data pipelines and to monitor, troubleshoot, and optimize cost and performance issues in accordance with best practices.
Questions
1124
Duration
130 minutes
Passing Score
720/1000
Difficulty
AssociateLast Updated
Jan 2026
The AWS Certified Data Engineer β Associate (DEA-C01) is an associate-level credential that validates a practitioner's ability to implement, monitor, and optimize data pipelines on AWS. Launched in 2023, it is the first AWS certification designed specifically for data engineers, replacing the need to cobble credentials from Solutions Architect or Data Analytics Specialty exams. The exam assesses proficiency across the full data engineering lifecycle: ingesting and transforming data, selecting and managing appropriate data stores, orchestrating pipelines using programming concepts, and enforcing data security and governance policies using AWS-native tooling.
Key AWS services in scope include Amazon S3, AWS Glue, Amazon Redshift, Amazon Kinesis, Amazon EMR, AWS Lake Formation, Amazon DynamoDB, AWS Database Migration Service, and Amazon Athena, among others. Candidates are evaluated on their ability to compare cost and performance trade-offs between services, apply SQL on AWS platforms, implement encryption and access controls, and validate data quality and consistency. Out-of-scope topics include ML model training and inference, programming-language-specific syntax, and deriving business conclusions from data analysis.
The target candidate is a data engineer or data architect with roughly 2β3 years of experience in data engineering and at least 1β2 years of hands-on AWS experience. This includes professionals who design and maintain ETL/ELT pipelines, manage data lakes and warehouses, or work with real-time streaming architectures. Adjacent roles transitioning into cloud data engineering β such as database administrators, backend developers, or traditional ETL developers β will also find this certification a clear roadmap for bridging legacy skills with AWS-native approaches.
The exam suits those who regularly work with concepts such as volume, variety, and velocity of data; data modeling and schema design; data lifecycle management; and cloud security and governance. It is not aimed at data scientists, ML engineers, or business analysts, as those domains fall outside the exam's scope.
AWS does not enforce formal prerequisites for the DEA-C01, but the official exam guide recommends 2β3 years of data engineering or data architecture experience and 1β2 years of hands-on work with AWS services. Candidates should be comfortable setting up and maintaining ETL pipelines from ingestion to destination, writing and executing SQL queries, using Git-based source control workflows, and applying language-agnostic programming concepts (loops, conditionals, data structures).
On the AWS side, recommended knowledge includes familiarity with data pipeline orchestration services (AWS Glue, AWS Step Functions), storage systems (Amazon S3, Amazon Redshift, Amazon DynamoDB), streaming platforms (Amazon Kinesis), and security/governance services (AWS IAM, AWS KMS, AWS Lake Formation). Understanding of data lakes, networking fundamentals (VPC, subnets, connectivity), compute options (Amazon EMR, AWS Lambda), and vector/embedding concepts is also beneficial. While no prior AWS certification is required, having the AWS Cloud Practitioner or AWS Solutions Architect β Associate background provides a useful foundation.
The DEA-C01 exam consists of 65 total questions: 50 scored questions that contribute to the final result and 15 unscored pilot questions that AWS uses to evaluate future content. Unscored questions are not identified, so candidates should treat all questions equally. Question types are multiple choice (one correct answer from four options) and multiple response (two or more correct answers from five or more options). The time limit is 130 minutes, and the exam is delivered via Pearson VUE at a testing center or through an online proctored session. The exam is available in English, Japanese, Korean, and Simplified Chinese, and costs $150 USD.
Scores are reported on a scaled range of 100β1,000, and the minimum passing score is 720. AWS uses a compensatory scoring model, meaning candidates do not need to achieve a passing threshold in each individual domain β only the overall scaled score matters. Unanswered questions are treated as incorrect; there is no penalty for guessing. The certification is valid for three years, after which recertification requires passing the current version of the exam.
The DEA-C01 certification targets one of the fastest-growing roles in cloud computing. AWS-certified data engineers in the US report average salaries around $141,000 per year according to Glassdoor data, with entry-level positions starting near $124,000β$130,000 and senior roles exceeding $175,000. Research from the Jefferson Frank Careers and Hiring Guide found that 73% of AWS professionals saw a salary increase after certification, averaging approximately 27%. Job roles accessible with this credential include Data Engineer, Cloud Data Architect, ETL/ELT Developer, Data Platform Engineer, and Analytics Engineer.
AWS certifications appear in cloud job postings more than any other vendor credential, and the DEA-C01 specifically validates the services β Glue, Redshift, Kinesis, S3 β that dominate real-world data engineering job requirements. For professionals transitioning from database administration, backend development, or traditional ETL roles, the certification provides a structured path into cloud-native data engineering. Many candidates report role transitions or salary increases within 3β6 months of earning the credential. Pairing DEA-C01 with the Databricks Data Engineer Associate certification is widely considered the most job-market-relevant two-certification combination in the data engineering space.
1. Adatum Insurance has a data pipeline that uses AWS Glue ETL to process claims data from a JDBC source (Amazon RDS MySQL). The job processes incrementally using job bookmarks. The data engineer notices that the job is reprocessing records it has already processed, causing duplicate entries in the target S3 location. Which combination of actions should the data engineer take to resolve this issue? (Select two!)
Select all that apply2. Tailspin Healthcare has an AWS Glue ETL job that processes medical records from multiple S3 sources. The source data contains nested JSON structures with fields that sometimes appear as strings and sometimes as integers across different source files. The job fails with type mismatch errors when attempting to write to the target table. What should the data engineer use to resolve the type ambiguities? (Select one!)
3. Fabrikam Analytics runs AWS Glue ETL jobs that read from a Glue Data Catalog table with over 2 million partitions stored in Amazon S3. The jobs filter on year, month, and day partition columns but take over 20 minutes just to list partitions before any data processing begins. Which combination of optimizations reduces the partition listing time with the LEAST operational overhead? (Select two!)
Select all that apply4. A data engineer at Contoso Corp is migrating a large Oracle database to Amazon Aurora PostgreSQL using AWS DMS. Several tables contain LOB columns with mixed sizes β most LOB values are under 10 KB, but approximately 5% exceed 100 KB. The engineer needs to maximize migration performance while ensuring no data is truncated. Which LOB handling mode should the engineer configure for the DMS task? (Select one!)
5. Litware Automotive uses Amazon Data Firehose to deliver vehicle telemetry data to Amazon S3. The data must be partitioned by vehicle_type and region extracted from the JSON payload. The team configures dynamic partitioning with JQ expressions but discovers that the minimum buffer interval is 60 seconds instead of the expected 0 seconds. What explains this behavior? (Select one!)
All exams included β’ Cancel anytime