Q: 3
An ML engineer has trained a neural network by using stochastic gradient descent (SGD). The neural
network performs poorly on the test set. The values for training loss and validation loss remain high
and show an oscillating pattern. The values decrease for a few epochs and then increase for a few
epochs before repeating the same cycle.
What should the ML engineer do to improve the training process?
Options
Discussion
Option C
Maybe D. The loss oscillates when the learning rate's too high so dropping it should help steady the model's updates. Not 100% but that's what I've seen in a few practice sets.
Yeah D tbh. High and oscillating loss is classic high learning rate. Dropping it should smooth things out, I think.
D imo. Had something like this in a mock exam, oscillating loss usually means learning rate is too high. Lower it to help stabilize training.
Be respectful. No spam.