Q: 7
You are managing a Slurm cluster with multiple GPU nodes, each equipped with different types of
GPUs. Some jobs are being allocated GPUs that should be reserved for other purposes, such as
display rendering.
How would you ensure that only the intended GPUs are allocated to jobs?
Options
Discussion
Makes sense to pick A. Direct config in gres.conf and slurm.conf is how GPU allocation is actually controlled with Slurm.
A. seeing this in recent exam reports too, configs through gres.conf are what actually restricts which GPUs Slurm uses.
Same, I'd pick A here. Only listing the right GPUs in gres.conf actually controls which ones Slurm will allocate. The other choices don’t prevent jobs from landing on display GPUs. Pretty sure that's what the exam wants.
Option A Official Slurm admin guide and most practice exams highlight gres.conf config for this. Labs emphasize it too.
A , official Slurm docs and admin labs push configuring gres.conf and slurm.conf for this exact scenario.
A tbh, but watch out-if gres.conf lists a GPU ID that’s swapped by the OS (like after reboot), Slurm might still allocate the wrong one. Seen that tripping people up in similar exam questions. Anyone disagree?
A imo, saw this on a similar practice exam. Official guide covers gres.conf usage.
Pretty sure it's A since Slurm relies on gres.conf to control which GPUs are visible for scheduling. Manual steps like nvidia-smi (B) or reinstalling drivers won’t restrict allocation, you have to exclude the display GPUs via config. Agree?
Is there any scenario where B would even work reliably? I always thought nvidia-smi assignments can't enforce job-level GPU isolation the way Slurm's gres.conf does. Seems like manual GPU allocation just isn't scalable for cluster setups.
A tbh, D feels like a trap since just increasing GPU requests doesn’t solve the config issue long-term. A is the standard way via gres.conf and slurm.conf control, but if anyone’s had success with another method let me know.
Be respectful. No spam.