I always thought it was B since you can pick master type, worker and parameter server counts, so B.
B is right here. compute.networkUser has to be on the service account running Dataflow, otherwise pipeline can't use the Shared VPC subnet. The agent role is for managing infra, not network access itself. Pretty sure that's the distinction but open if anyone sees it another way.
I don't think it's A. B fits better here since Cloud Pub/Sub is designed for unreliable or lossy networks and saves you from Kafka maintenance overhead. A is tempting but adds cost and complexity you don't need for the actual requirement. Pretty sure B's the expected answer on these types of Google questions.
Is anyone using the official study guide or practice tests to prep for these scenario questions? I keep seeing similar cases where Pub/Sub is the recommended approach for flaky connections. Just want to confirm if the guide mentions this pattern clearly.
A and B are what I'd pick here. Scaling worker count or boosting instance type directly tackles CPU starvation in Dataflow. The buffer options (D, E) deal more with throughput spikes or persistent queueing, not pure compute limits. Pretty sure this is the intent, but open if anyone disagrees.
Had something like this in a mock before. Isn’t increasing max workers (A) and using bigger instances (B) the straightforward way to deal with CPU bottlenecks? Adding buffers like D or E sounds more for handling message surges, not compute limits. Agree?
I don't think D or E make sense here, since those buffer options just add complexity and won't solve the CPU bottleneck. A and B are way more direct, either scale out or use beefier machines. C's tricky but violates EU data residency, so that's probably a trap. Agree?
C . Had something like this in a mock and Google's guidance is to materialize dimensions using views when you need joins in a star schema, especially if you want to speed things up but not use more storage. Partitioning would help for date filters, but the question asks about storage impact specifically. Might be tricky but I’d stick with C here. Agree?
Option B is correct. When your endpoint has an out-of-date SSL certificate, Cloud Pub/Sub can't confirm delivery due to failed TLS handshake, so it keeps retrying and you get duplicate messages. Option D is tempting since missing acks also trigger retries, but this scenario is specific to SSL issues with push endpoints. Seen similar guidance in Google docs. Open to other takes though if anyone's seen differently!