-
Hi, I'm interested in using Build a Large Language Model (From Scratch) for a school project, but I’d like to make sure my PC specs are sufficient before purchasing the book. Here are my specs: CPU: 12th Gen Intel Core i7-12650H @ 2.30 GHz RAM: 32 GB GPU: NVIDIA GeForce RTX 3050 Laptop GPU (6GB VRAM) Available storage: ~150 GB Would this setup be enough to follow the practical parts of the book, at least for training small-scale models or running examples locally? Thank you for your help! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
Hi, Imo, on CPU training with your ram, you'll have no problem, you'll be able to run all chapters smoothly without having to worry about memory. Training on CPU is a bit slow though. If instead you want to use your Cuda GPU, some parts might lead you into OOM territory in f32, depending on the chapters, not sure. Sebastian made the book with these limits and a broader audience in mind, I don't think you'll run into trouble tbh. |
Beta Was this translation helpful? Give feedback.
Hi,
Imo, on CPU training with your ram, you'll have no problem, you'll be able to run all chapters smoothly without having to worry about memory. Training on CPU is a bit slow though.
If instead you want to use your Cuda GPU, some parts might lead you into OOM territory in f32, depending on the chapters, not sure.
@rasbt can give a better answer on that, I don't remember exactly. But It'll be feasible in bf16
model.to(torch.bfloat16)
.You can also reduce some hyperparameters like
batch_size
orcontext_length
to fit everything in VRAM. Even use Google Colab too, if needed.Sebastian made the book with these limits and a broader audience in mind, I don't think you'll run into trouble tbh.