Q: 1
You work as an analyst at a large banking firm. You are developing a robust, scalable ML pipeline to
train several regression and classification models. Your primary focus for the pipeline is model
interpretability. You want to productionize the pipeline as quickly as possible What should you do?
Options
Discussion
B tbh, GKE with XGBoost custom training sounds scalable and lets you fine-tune stuff, which I thought is good for productionizing quickly. Plus XGBoost gives some model interpretability (like feature importance). Not 100% sure though, D might be more purpose-built for orchestration. Anyone see issues with B?
Why not just use A if you want to skip custom orchestration steps? Cloud Composer feels heavier for pure speed.
Pretty sure it's D. Official exam guide and practice questions both recommend Cloud Composer for pipeline orchestration and model interpretability.
Its D, seen similar question in a mock and Cloud Composer fits this use case.
Be respectful. No spam.
Q: 2
You work for a bank You have been asked to develop an ML model that will support loan application
decisions. You need to determine which Vertex Al services to include in the workflow You want to
track the model's training parameters and the metrics per training epoch. You plan to compare the
performance of each version of the model to determine the best model based on your chosen
metrics. Which Vertex Al services should you use?
Options
Discussion
Option C is right here. ML Metadata logs the artifacts, Experiments helps with model version comparisons, and TensorBoard shows metrics per epoch. Pretty sure this trio covers exactly what's needed for tracking and evaluation.
B or D here. I figure Pipelines helps with end-to-end workflow tracking and Experiments (in B) or TensorBoard (in D) let you compare model versions and metrics. Not totally sure if ML Metadata is really required for tracking training parameters, though. Anyone disagree?
I'm not fully sure here, but think it's C. These sound like the tools for tracking experiments and training details? Can someone confirm if that's correct?
Be respectful. No spam.
Q: 3
You developed a custom model by using Vertex Al to forecast the sales of your company s products
based on historical transactional data You anticipate changes in the feature distributions and the
correlations between the features in the near future You also expect to receive a large volume of
prediction requests You plan to use Vertex Al Model Monitoring for drift detection and you want to
minimize the cost. What should you do?
Options
Discussion
Probably D. You want both features and attributions for more insight into drift, but keeping that prediction-sampling-rate down (closer to 0) is what really helps minimize cost as usage scales. Pretty sure that's the balanced approach if you want visibility but need to watch spend. Disagree?
B seems like a good pick because setting the prediction-sampling-rate closer to 1 should catch more drift cases with lots of data, which matters if you expect major changes. Plus, only monitoring features feels simpler and reduces complexity so might save costs too. Maybe missing something about feature attributions though-open to being corrected.
Be respectful. No spam.
Q: 4
You work for a pet food company that manages an online forum Customers upload photos of their
pets on the forum to share with others About 20 photos are uploaded daily You want to
automatically and in near real time detect whether each uploaded photo has an animal You want to
prioritize time and minimize cost of your application development and deployment What should you
do?
Options
Discussion
I don’t think D is right. A is better since Cloud Vision API gives you object localization out of the box, no need to train or label datasets. It’s faster and cheaper for a simple animal/not animal check, pretty sure that’s what they want.
B. not D. Vendor practice tests cover Cloud Vision API usage like this pretty often.
A , D is a trap since manual labeling and training takes longer and costs more for this use case.
A makes sense for speed and cost. Cloud Vision API is ready to use, cuts out training and manual labeling. Practice with the official GCP console helps see how fast it works.
Cloud Vision API is probably the fastest and easiest here, so I'd pick A. Official docs and some hands-on with Vision API in the console are super helpful for this type of real-time detection use case. Anyone see a downside?
Be respectful. No spam.
Q: 5
You recently developed a deep learning model using Keras, and now you are experimenting with
different training strategies. First, you trained the model using a single GPU, but the training process
was too slow. Next, you distributed the training across 4 GPUs using tf.distribute.MirroredStrategy
(with no other changes), but you did not observe a decrease in training time. What should you do?
Options
Discussion
Probably D. If you just switch to MirroredStrategy but keep the same (small) batch size, each GPU isn't used efficiently so no real speed gain. Increasing the batch size lets GPUs process more data in parallel. Not 100% sure if dataset sharding is a trap here, but D is what I see in similar questions.
Is the question assuming we want to keep model accuracy exactly the same, or is a minor change in accuracy acceptable if we can finish training faster? That would make D a safer pick.
Be respectful. No spam.
Q: 6
You need to train a computer vision model that predicts the type of government ID present in a given
image using a GPU-powered virtual machine on Compute Engine. You use the following parameters:
• Optimizer: SGD
• Image shape 224x224
• Batch size 64
• Epochs 10
• Verbose 2
During training you encounter the following error: ResourceExhaustedError: out of Memory (oom)
when allocating tensor. What should you do?
Options
Discussion
B tbh, since batch size eats up a lot of GPU memory fast. D might be tempting but you'd lose image detail for IDs, so not ideal here. Saw a similar question in practice and B was correct. Trap is thinking optimizer or learning rate helps!
Reducing batch size (B) is usually the first thing to try for a ResourceExhaustedError since it scales down tensor allocations pretty quickly. Lowering image shape (D) works too but risks losing critical features in ID images. Pretty sure B is expected, unless resolution drops are acceptable. Agree?
Why not D? Smaller image shape means less memory per input, might solve OOM too.
Guessing B here. Batch size directly impacts GPU memory usage, so that's usually the first thing to drop if you see ResourceExhaustedError. Saw a similar question on a practice exam and B was right. Anyone disagree?
Its B, batch size directly impacts GPU memory use. Seen this error before, lowering batch size fixes it fast. Agree?
Be respectful. No spam.
Q: 7
Your data science team has requested a system that supports scheduled model retraining, Docker
containers, and a service that supports autoscaling and monitoring for online prediction requests.
Which platform components should you choose for this system?
Options
Discussion
B . Vertex AI Pipelines does scheduled retraining, Vertex AI Prediction covers online predictions with autoscaling, and Model Monitoring checks the deployed models. Not totally sure but pretty sure the others miss something here. Can someone confirm?
B is the only one that covers everything, with Pipelines for retraining and Prediction plus Model Monitoring to hit autoscaling and observability requirements. App Engine doesn't fit the container or prediction piece. Pretty sure this is what Google wants here.
Be respectful. No spam.
Q: 8
Your team has a model deployed to a Vertex Al endpoint You have created a Vertex Al pipeline that
automates the model training process and is triggered by a Cloud Function. You need to prioritize
keeping the model up-to-date, but also minimize retraining costs. How should you configure
retraining'?
Options
Discussion
Option D not B. Only D triggers on real feature drift so you don't retrain for no reason.
D imo, feature drift monitoring is best to avoid unnecessary retrains. B might look cheaper but doesn't account for when the model actually needs updating. Seen similar on practice tests, pretty sure D is what they want.
C vs D? Saw similar question on practice, both seem valid but C feels right since anomaly could trigger retrain too.
C/D? Both use anomaly detection but C doesn't explicitly mention drift, so not totally sure.
Nice question, really clear scenario. B
Be respectful. No spam.
Q: 9
You are developing a Kubeflow pipeline on Google Kubernetes Engine. The first step in the pipeline is
to issue a query against BigQuery. You plan to use the results of that query as the input to the next
step in your pipeline. You want to achieve this in the easiest way possible. What should you do?
Options
Discussion
D here, but honestly C looks ok too if you prefer custom stuff. Pretty sure D is quicker for most cases though.
Its C for me. Using the DSL with the Python BigQuery client lets you customize more and keep things in code, which I thought was best practice. But maybe D is simpler if you don't need anything special. Anyone think that's totally wrong?
C/D? Not sure honestly, both would do the job but D is probably easier since it's ready-made. Anyone disagree with that?
D imo, saw a similar question in some exam reports and the pre-built BigQuery component is the fastest way. No need to reinvent, just plug it in and go. Let me know if you think otherwise.
Be respectful. No spam.
Q: 10
You are experimenting with a built-in distributed XGBoost model in Vertex AI Workbench user-
managed notebooks. You use BigQuery to split your data into training and validation sets using the
following queries:
CREATE OR REPLACE TABLE ‘myproject.mydataset.training‘ AS
(SELECT * FROM ‘myproject.mydataset.mytable‘ WHERE RAND() < 0.8);
CREATE OR REPLACE TABLE ‘myproject.mydataset.validation‘ AS
(SELECT * FROM ‘myproject.mydataset.mytable‘ WHERE RAND() < 0.2);
After training the model, you achieve an area under the receiver operating characteristic curve (AUC
ROC) value of 0.8, but after deploying the model to production, you notice that your model
performance has dropped to an AUC ROC value of 0.65. What problem is most likely occurring?
Options
Discussion
A is wrong, C looks right. The issue is the RAND() split allows overlap between training and validation sets, so you evaluate on data already seen by the model, which hides real-world generalization problems. Pretty sure that’s what’s going on here.
C imo, seen similar scenario in practice sets and the question is clear about overlapping splits.
Be respectful. No spam.
Question 1 of 20 · Page 1 / 2