MLflow PyFunc is the go-to method here, so D. Saw this approach recommended in the official Databricks guide. If anyone saw another method used in recent exam practice, let me know.
I don’t think C is the way. Preprocessing the prompt up front usually affects how the LLM interprets your inputs, which can make a bigger difference than cleaning up the output afterward. Postprocessing helps but doesn’t fix issues caused by a poorly structured prompt. Pretty sure D is more standard but open to other takes.
Yeah, D looks right to me. Using an MLflow PyFunc model gives you a clean way to bundle custom preprocessing steps with LLM calls, which is super handy for production pipelines in Databricks. Directly modifying the LLM architecture (A) is risky and not typical here. I think D but happy to be challenged if someone found a better approach.
D is correct for Databricks flow. If the question asked for the most secure method, would that change the pick?