Q: 11
You work at a subscription-based company. You have trained an ensemble of trees and neural
networks to predict customer churn, which is the likelihood that customers will not renew their
yearly subscription. The average prediction is a 15% churn rate, but for a particular customer the
model predicts that they are 70% likely to churn. The customer has a product usage history of 30%, is
located in New York City, and became a customer in 1997. You need to explain the difference
between the actual prediction, a 70% churn rate, and the average prediction. You want to use Vertex
Explainable AI. What should you do?
Options
Discussion
B . Some folks go for C but integrated gradients is mostly for images and text, not tabular data like churn. Easy trap there.
B
I don’t think it’s B. C. Integrated gradients sometimes get used for explainability, especially when you want to trace prediction changes as inputs shift from baseline, and might still help even if it’s not image data. Maybe there’s a trap here with Shapley?
Probably B, since Vertex Explainable AI's sampled Shapley explanations break down the prediction difference by feature. Makes sense for tabular data like this (churn, usage, location). Integrated gradients are more for images or text, so not a fit here. Pretty sure this matches what Google recommends. Disagree?
B tbh, because sampled Shapley in Vertex Explainable AI is the way to show individual feature contributions for tabular models. Confident pick.
Be respectful. No spam.
Q: 12
You are developing ML models with Al Platform for image segmentation on CT scans. You frequently
update your model architectures based on the newest available research papers, and have to rerun
training on the same dataset to benchmark their performance. You want to minimize computation
costs and manual intervention while having version control for your code. What should you do?
Options
Discussion
C/D? Cloud Composer (D) feels like a trap here since it polls on a schedule and doesn't handle versioning well. C lets you automate with repo triggers and proper code management. Pretty sure it's C but let me know if you think Composer makes sense for this use case.
Option D
Be respectful. No spam.
Q: 13
You work for the AI team of an automobile company, and you are developing a visual defect
detection model using TensorFlow and Keras. To improve your model performance, you want to
incorporate some image augmentation functions such as translation, cropping, and contrast
tweaking. You randomly apply these functions to each training batch. You want to optimize your data
processing pipeline for run time and compute resources utilization. What should you do?
Options
Discussion
Probably B since Keras generators can handle augmentations and batching together, saw similar approach in some exam reports.
Be respectful. No spam.
Q: 14
You recently deployed a scikit-learn model to a Vertex Al endpoint You are now testing the model on
live production traffic While monitoring the endpoint. you discover twice as many requests per hour
than expected throughout the day You want the endpoint to efficiently scale when the demand
increases in the future to prevent users from experiencing high latency What should you do?
Options
Discussion
B is the way to go. Setting minReplicaCount helps Vertex AI autoscale properly for spikes without risking high latency. The other choices don't directly address scaling for unpredictable traffic. Pretty sure about this, but open to corrections if I'm missing something.
Be respectful. No spam.
Q: 15
You developed a custom model by using Vertex Al to predict your application's user churn rate You
are using Vertex Al Model Monitoring for skew detection The training data stored in BigQuery
contains two sets of features - demographic and behavioral You later discover that two separate
models trained on each set perform better than the original model
You need to configure a new model mentioning pipeline that splits traffic among the two models You
want to use the same prediction-sampling-rate and monitoring-frequency for each model You also
want to minimize management effort What should you do?
Options
Discussion
If the "minimize management effort" part wasn't required, would C make more sense than D here?
Be respectful. No spam.
Q: 16
You need to develop an image classification model by using a large dataset that contains labeled
images in a Cloud Storage Bucket. What should you do?
Options
Discussion
Its C if "best" means lowest effort, but is there a code requirement that would make A or B more suitable?
I don’t think it’s C. B fits better if you have to use TensorFlow Extended for custom preprocessing or model logic.
Be respectful. No spam.
Q: 17
You have created a Vertex Al pipeline that includes two steps. The first step preprocesses 10 TB data
completes in about 1 hour, and saves the result in a Cloud Storage bucket The second step uses the
processed data to train a model You need to update the model's code to allow you to test different
algorithms You want to reduce pipeline execution time and cost, while also minimizing pipeline
changes What should you do?
Options
Discussion
Not sure about that, pretty sure A is better here. If you add a pipeline parameter and step to decide whether to preprocess or not, you can skip preprocessing if data hasn't changed and just go straight to training. That way, you're not redoing heavy ETL work every time, and the change is minimal. But open to other ideas if I'm missing something obvious?
I'd pick C here. More CPU and RAM for the preprocessing step should speed things up without changing the pipeline much, I think. Disagree?
D (I'm a bit confused but caching seems to save rerunning preprocessing, which cuts cost and keeps things simple?)
Be respectful. No spam.
Q: 18
You are training an ML model on a large dataset. You are using a TPU to accelerate the training
process You notice that the training process is taking longer than expected. You discover that the TPU
is not reaching its full capacity. What should you do?
Options
Discussion
Can't believe Google still makes us guess at this stuff. Its D because TPUs are designed to chew through big batches, so if utilization is low, the batch size is probably too small. Bumping it up lets you push more data per step and make better use of the parallel cores. I think that's the main idea here, unless I'm missing something.
Not A, D. Larger batch size helps TPUs handle more data each step, so you can max out the hardware.
Be respectful. No spam.
Q: 19
You are an ML engineer at a bank. You have developed a binary classification model using AutoML
Tables to predict whether a customer will make loan payments on time. The output is used to
approve or reject loan requests. One customer’s loan request has been rejected by your model, and
the bank’s risks department is asking you to provide the reasons that contributed to the model’s
decision. What should you do?
Options
Discussion
Its A. Had something like this in a mock before and local feature importance is what they want for a specific prediction.
Not C, A is right here, but is the question asking for the specific reasons for this one customer's decision or general model behavior? If they wanted global (whole model) insights, C could look tempting.
Be respectful. No spam.
Q: 20
You are working on a Neural Network-based project. The dataset provided to you has columns with
different ranges. While preparing the data for model training, you discover that gradient
optimization is having difficulty moving weights to a good solution. What should you do?
Options
Discussion
Its A, but if the question said 'best first step,' I'd rethink for B.
Be respectful. No spam.
Question 11 of 20 · Page 2 / 2