Q: 13
A Generative Al Engineer is building a system that will answer questions on currently unfolding news
topics. As such, it pulls information from a variety of sources including articles and social media
posts. They are concerned about toxic posts on social media causing toxic outputs from their system.
Which guardrail will limit toxic outputs?
Options
Discussion
Option A was on my real exam. Limiting input sources to approved accounts is the core guardrail here since it stops toxic data before it even gets processed by the LLM. Pretty sure A is right, unless I'm missing something.
A , similar questions in the official guide focus on input filtering as the main guardrail. Blocking toxic sources up front is way more effective than just logging outputs or adding rate limits. Not completely sure if there's a newer best practice but all the recent practice sets lean toward A.
A here. Filtering inputs at the source comes before logging or rate limiting, so it actually blocks toxic content from even reaching the LLM. D looks like a trap since monthly checks are too late to prevent issues up front.
Probably A. Only letting data from vetted sources through helps block toxic stuff before it ever gets to the LLM. Logging or rate limiting doesn't really prevent bad content in real time. Pretty sure A matches the usual best practice, agree?
Could flip to D if the requirement was monitoring not prevention, but for actual guardrails it’s A.
Its D. If we log all the LLM outputs and do a monthly toxicity analysis, we can catch any harmful trends and address them in batches. I think this way gives us historical insight, so it feels more robust to me. Disagree?
Be respectful. No spam.