Scenario: A Bedrock chatbot uses Amazon Titan Text but provides generic answers because it lacks access to proprietary order management and product documentation data (S3, internal DB). The team needs to enhance responses using this private data without retraining the model. Question- Which option satisfies this requirement?.. Options:
Scenario: A team needs to fine-tune an LLM for text summarization using a low- code/no-code (LCNC) solution to automate model training and minimize manual intervention. Question- Which solution will best meet the team’s requirements?. Options:
Probably D. Only D mentions using Titan Embeddings from Bedrock and OpenSearch vectors, which is the actual pattern for native RAG workflows on AWS now. Kendra does semantic search but doesn't expose embeddings for external use with agents, that's the trap here. Pretty sure D fits best; open to pushback if there's a Kendra feature I missed.
Pretty clear it's A for this scenario. The question asks specifically for a low-code/no-code way to automate, and JumpStart with Autopilot is built exactly for that-minimal manual work, mostly just point and click. B and D need way more hands-on scripting or setup, so they miss the automation piece. I think A covers what AWS expects here, but open to counterpoints if I'm missing something subtle.
Yeah, A fits best here since it puts all the low-code/no-code and automation requirements front and center. JumpStart plus Autopilot means you mostly click through a UI without much manual setup. B and D let you control more, but that’s not really what the question wants. Pretty sure A is what exam writers want for LCNC stuff, but let me know if I’m missing something.
A for sure, since Autopilot with JumpStart nails the LCNC angle and keeps everything mostly UI-driven. The other options want more manual steps or script work, which doesn't fit the "automation/minimal intervention" part. Pretty sure this is what AWS wants for anything LCNC-focused, but open if there's a catch I'm missing.
Scenario: An AI developer needs to systematically determine how varying PySpark feature transformation parameters and sample sizes affects overall model accuracy and inference performance. Question- Which solution will meet this requirement most effectively?. Options:
Had something like this in a mock. For real LCNC, D's the only one that gives both JumpStart and Autopilot together, so you barely need to touch code and the process is automated. The other options involve more manual work or require scripts. Pretty confident with D here but open if anyone's seen AWS change that recently.
Scenario: An administrator must ensure that each AI developer can access only their assigned SageMaker notebook instance while maintaining shared access to Amazon Rekognition APIs and training data stored in Amazon S3. Question- Which solution will meet this requirement?. Options:
Scenario: An image classifier misclassifies images, and analysis shows the model is highly sensitive to image orientation (e.g., upside-down pandas). The team needs to enhance the model's ability to identify the objects regardless of orientation without collecting a new dataset. Question- Which approach most effectively enhances the model’s accuracy in addressing this specific misclassification issue?. Options: