Q: 9
You are developing a Kubeflow pipeline on Google Kubernetes Engine. The first step in the pipeline is
to issue a query against BigQuery. You plan to use the results of that query as the input to the next
step in your pipeline. You want to achieve this in the easiest way possible. What should you do?
Options
Discussion
D here, but honestly C looks ok too if you prefer custom stuff. Pretty sure D is quicker for most cases though.
D
saw pretty similar problem in my exam, in a practice exam, D is what they wanted there too.
Option D makes sense since it's literally designed for this. The Kubeflow prebuilt BigQuery Query Component saves coding time and works out of the box with pipelines. I get why some might pick C, but that's extra overhead for no reason here. Pretty sure D is the easiest if that's what they're asking for.
Nah, I don't think it's C this time. D is the prebuilt component trap wins here.
Isn't this a classic case where just checking the official Kubeflow docs or practice exams clears it up?
D imo. Using a prebuilt BigQuery Query Component saves time and avoids extra Python coding or Docker image building. If the question was about flexibility instead, maybe C would fit, but "easiest" points to D here. Not 100 percent sure though, could see an argument for C in restricted environments.
B or C. If the Kubeflow registry is blocked or needs custom query logic, writing a Python script (B) or making your own component (C) could be easier in practice. Not sure which one is actually faster.
Not sure, but shouldn't we just go with the built-in Kubeflow component for this instead of coding from scratch?
B tbh
Be respectful. No spam.