Multi-GPU Training - Two Suggested Improvements #11079
Replies: 1 comment
-
Hello!
The answer is yes: The whole purpose of the LightningModule is to interact with the Trainer. It exposes some hooks like If you are not sure whether to use/convert Lightning or not, or are overwhelmed by the Trainer, we also have a lightweight version called LightningLite. It only bundles the accelerators (multi-GPU, TPU etc.) and comes without a Trainer, so that you can use your existing training loop and nn.Modules.
There could be multiple reasons, the most likely that you have not configured the Trainer correctly. Try
You should see a message printed like this when you run the script.
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi! I'm not sure whether this is appropriate to open as an issue, but I can't find answers to two key multi-GPU training questions:
Does one need to use Trainer in order to take advantage of PL's multi-GPU training? Or is the trainer unnecessary if the models are LightningModules?
If I look at nvidia-smi and see that my PL code isn't using GPUs, how do I identify the cause?
cc @Borda
Beta Was this translation helpful? Give feedback.
All reactions