Q: 2
When selecting parameters to optimize a prompt-tuned model experiment in IBM watsonx, which
parameter is the most critical for controlling the model’s ability to generate coherent and contextually
accurate responses?
Options
Discussion
C imo
Definitely learning rate here, so C. That parameter pretty much decides if prompt tuning actually helps the model learn to generate sensible, context-aware responses. Lab work and IBM docs both focus on tuning learning rate for this reason. Not 100 percent but that's how I've seen it explained in official guides-anyone see something different on real exams?
Learning rate is what matters most for coherent and context-aware responses, not batch size here. C.
Option B Had something like this in a mock, picked batch size for model quality.
Its B, batch size. Saw a similar question in practice that focused on batch size impacting learning dynamics.
Be respectful. No spam.