I don't think user-generated content (A) is the best pick here. D (Benchmark datasets) makes more sense since they're already labeled for bias and widely used for standard evaluation, so you skip all the manual setup and data cleaning. Moderation logs or guidelines would take way more admin work to adapt. Anyone disagree?
Q: 1
A social media company wants to use a large language model (LLM) for content moderation. The
company wants to evaluate the LLM outputs for bias and potential discrimination against specific
groups or individuals.
Which data source should the company use to evaluate the LLM outputs with the LEAST
administrative effort?
Options
Discussion
D for sure. Benchmark datasets are ready-to-go and already labeled for bias, so you don't need to set up your own framework or gather/annotate raw user data. Way less admin effort compared to using logs or starting from scratch with guidelines. Pretty confident here but open if someone sees it differently.
Probably D, benchmark datasets are already set up for this. Nice straightforward question.
Be respectful. No spam.
Q: 2
A company wants to create a chatbot by using a foundation model (FM) on Amazon Bedrock. The FM
needs to access encrypted data that is stored in an Amazon S3 bucket.
The data is encrypted with Amazon S3 managed keys (SSE-S3).
The FM encounters a failure when attempting to access the S3 bucket data.
Which solution will meet these requirements?
Options
Discussion
Its A. The big issue here is really IAM permissions for Bedrock's assumed role to decrypt S3 data, especially with SSE-S3 encryption. Saw similar in AWS docs and official exams, so pretty confident.
C or D? If the question said "most secure" or required sensitive data to stay protected, I'd pick differently, since C feels like a trap if permissions are the main blocker.
Be respectful. No spam.
Q: 3
A company is building an ML model. The company collected new data and analyzed the data by
creating a correlation matrix, calculating statistics, and visualizing the data.
Which stage of the ML pipeline is the company currently in?
Options
Discussion
This looks super close to one I had on a mock, the correlation matrix part basically nails it as C.
Skip B, C here. Calculating stats and making correlation matrices is classic EDA, not engineering features. Trap answer is B.
C or B-if "analyzed" also means creating new features, would that count as feature engineering instead? Just want to be sure on what they mean by 'analyzed.'
B tbh. . Saw a similar scenario in the official practice test, so I'd check the AWS study guide and practice banks for how these stages are defined.
All those activities scream exploratory data analysis to me, so C is right. Making correlation matrices and visualizations isn't creating new features. Pretty sure this is what AWS expects.
Similar question was in the official practice, recommend reviewing the AWS study guide for pipeline stages.
Be respectful. No spam.
Q: 4
A company has petabytes of unlabeled customer data to use for an advertisement campaign. The
company wants to classify its customers into tiers to advertise and promote the company's products.
Which methodology should the company use to meet these requirements?
Options
Discussion
B tbh
A is out, B makes sense if the customer data is really unlabeled. Classic clustering scenario.
Probably B, clustering fits when you don't have labeled data. Seen this kind of scenario before.
A is wrong, B. Unsupervised learning is what you want for tons of unlabeled data. Pretty sure saw similar logic in official practice exams and the AWS docs.
Its B, had something like this in a mock. Unlabeled data means unsupervised learning.
Be respectful. No spam.
Q: 5
Which technique breaks a complex task into smaller subtasks that are sent sequentially to a large
language model (LLM)?
Options
Discussion
Option D RAG involves bringing in external info to help answer questions, so I think it's about enriching the model's knowledge, not breaking up tasks. Sequential subtasks sounds more like how RAG retrieves multiple docs step by step. Not 100% sure though, open to corrections.
Be respectful. No spam.
Q: 6
Which option describes embeddings in the context of AI?
Options
Discussion
Its D, but people mix this up since visualizations like t-SNE use embeddings after the fact.
C or D? I've seen embeddings used for visualizing high-dimensional data (t-SNE, PCA), not just reducing dimensions. Anyone else think C fits?
Be respectful. No spam.
Q: 7
An AI company periodically evaluates its systems and processes with the help of independent
software vendors (ISVs). The company needs to receive email message notifications when an ISV's
compliance reports become available.
Which AWS service can the company use to meet this requirement?
Options
Discussion
B . AWS Artifact is where compliance reports usually live, so I figured that's the place for notifications on new reports. Data Exchange feels more about datasets than compliance docs. Anyone else picked B on similar questions?
C/D? Compliance checking kinda feels like Trusted Advisor too.
D imo, since AWS Data Exchange lets companies subscribe to datasets from third-party providers like ISVs. When a new compliance report is published, you can set up notifications via EventBridge to SNS/email. Audit Manager and Artifact are more for internal or AWS-specific reporting. Not 100% sure but that's how it's usually done in similar exam questions. Anyone seen this behave differently in labs or practice?
Be respectful. No spam.
Q: 8
A customer service team is developing an application to analyze customer feedback and
automatically classify the feedback into different categories. The categories include product quality,
customer service, and delivery experience.
Which AI concept does this scenario present?
Options
Discussion
B is the one you want here. Since it's written customer feedback, that's text data, which means natural language processing (NLP) is what classifies it. Computer vision (A) would only fit if we were dealing with images, not text. Pretty sure about this but let me know if there's another angle.
Its B for sure. Classifying written feedback is classic NLP, since it’s all about handling and understanding text data. Official AWS study guides and doing a few hands-on labs with SageMaker NLP would help nail this concept if you’re still unsure.
Option C. I figured recommendation systems help sort stuff into categories for users based on patterns. Not super confident though, since the question mentioned classification not suggestions. Let me know if I'm off.
B imo. But if the feedback was images or voice instead, would that change it to something else like A?
Be respectful. No spam.
Q: 9
A company manually reviews all submitted resumes in PDF format. As the company grows, the
company expects the volume of resumes to exceed the company's review capacity. The company
needs an automated system to convert the PDF resumes into plain text format for additional
processing.
Which AWS service meets this requirement?
Options
Discussion
B tbh is a trap, that's for personalization not text extraction. A is the only one that handles PDFs directly.
Probably A, saw a almost identical question in a practice set and Textract is the AWS tool for extracting text from PDFs.
Be respectful. No spam.
Q: 10
A company trained an ML model on Amazon SageMaker to predict customer credit risk. The model
shows 90% recall on training data and 40% recall on unseen testing data.
Which conclusion can the company draw from these results?
Options
Discussion
A is right. Huge drop in recall tells me the model memorized patterns instead of generalizing, classic overfit scenario. More data (C) would help, but the main issue here is overfitting for sure. Agree?
Probably C, lack of enough training data could explain the drop in recall on new data.
Be respectful. No spam.
Question 1 of 20 · Page 1 / 2