Q: 11
A company is building a data stream processing application. The application runs in an Amazon
Elastic Kubernetes Service (Amazon EKS) cluster. The application stores processed data in an Amazon
DynamoDB table.
The company needs the application containers in the EKS cluster to have secure access to the
DynamoDB table. The company does not want to embed AWS credentials in the containers.
Which solution will meet these requirements?
Options
Discussion
B is correct here. IRSA lets your EKS pods assume IAM roles without hardcoding credentials, which is way more secure than passing keys around. Nice clear scenario, seen similar ones in practice tests!
Be respectful. No spam.
Q: 12
A company uses an Amazon QuickSight dashboard to monitor usage of one of the company's
applications. The company uses AWS Glue jobs to process data for the dashboard. The company
stores the data in a single Amazon S3 bucket. The company adds new data every day.
A data engineer discovers that dashboard queries are becoming slower over time. The data engineer
determines that the root cause of the slowing queries is long-running AWS Glue jobs.
Which actions should the data engineer take to improve the performance of the AWS Glue jobs?
(Choose two.)
Options
Discussion
C
Be respectful. No spam.
Q: 13
A retail company has a customer data hub in an Amazon S3 bucket. Employees from many countries
use the data hub to support company-wide analytics. A governance team must ensure that the
company's data analysts can access data only for customers who are within the same country as the
analysts.
Which solution will meet these requirements with the LEAST operational effort?
Options
Discussion
No comments yet. Be the first to comment.
Be respectful. No spam.
Q: 14
A data engineer is building a data pipeline on AWS by using AWS Glue extract, transform, and load
(ETL) jobs. The data engineer needs to process data from Amazon RDS and MongoDB, perform
transformations, and load the transformed data into Amazon Redshift for analytics. The data updates
must occur every hour.
Which combination of tasks will meet these requirements with the LEAST operational overhead?
(Choose two.)
Options
Discussion
No comments yet. Be the first to comment.
Be respectful. No spam.
Q: 15
A company's data engineer needs to optimize the performance of table SQL queries. The company
stores data in an Amazon Redshift cluster. The data engineer cannot increase the size of the cluster
because of budget constraints.
The company stores the data in multiple tables and loads the data by using the EVEN distribution
style. Some tables are hundreds of gigabytes in size. Other tables are less than 10 MB in size.
Which solution will meet these requirements?
Options
Discussion
No comments yet. Be the first to comment.
Be respectful. No spam.
Q: 16
A company uses Amazon Redshift for its data warehouse. The company must automate refresh
schedules for Amazon Redshift materialized views.
Which solution will meet this requirement with the LEAST effort?
Options
Discussion
B, not C
Be respectful. No spam.
Q: 17
A retail company stores order information in an Amazon Aurora table named Orders. The company
needs to create operational reports from the Orders table with minimal latency. The Orders table
contains billions of rows, and over 100,000 transactions can occur each second.
A marketing team needs to join the Orders data with an Amazon Redshift table named Campaigns in
the marketing team's data warehouse. The operational Aurora database must not be affected.
Which solution will meet these requirements with the LEAST operational effort?
Options
Discussion
No comments yet. Be the first to comment.
Be respectful. No spam.
Q: 18
A company currently uses a provisioned Amazon EMR cluster that includes general purpose Amazon
EC2 instances. The EMR cluster uses EMR managed scaling between one to five task nodes for the
company's long-running Apache Spark extract, transform, and load (ETL) job. The company runs the
ETL job every day.
When the company runs the ETL job, the EMR cluster quickly scales up to five nodes. The EMR cluster
often reaches maximum CPU usage, but the memory usage remains under 30%.
The company wants to modify the EMR cluster configuration to reduce the EMR costs to run the daily
ETL job.
Which solution will meet these requirements MOST cost-effectively?
Options
Discussion
No comments yet. Be the first to comment.
Be respectful. No spam.
Q: 19
A company currently stores all of its data in Amazon S3 by using the S3 Standard storage class.
A data engineer examined data access patterns to identify trends. During the first 6 months, most
data files are accessed several times each day. Between 6 months and 2 years, most data files are
accessed once or twice each month. After 2 years, data files are accessed only once or twice each
year.
The data engineer needs to use an S3 Lifecycle policy to develop new data storage rules. The new
storage solution must continue to provide high availability.
Which solution will meet these requirements in the MOST cost-effective way?
Options
Discussion
No comments yet. Be the first to comment.
Be respectful. No spam.
Q: 20
A company needs to load customer data that comes from a third party into an Amazon Redshift data
warehouse. The company stores order data and product data in the same data warehouse. The
company wants to use the combined dataset to identify potential new customers.
A data engineer notices that one of the fields in the source data includes values that are in JSON
format.
How should the data engineer load the JSON data into the data warehouse with the LEAST effort?
Options
Discussion
B. seen similar in practice tests.
Be respectful. No spam.
Question 11 of 20 · Page 2 / 2