Skip to content

Finetune vs Lora settings. (need some help) #251

Closed Answered by Calamdor
Z1-Gamer asked this question in Q&A
Discussion options

You must be logged in to vote

Generally speaking, updating the gradients during a finetune is a destructive event. A LoRA at batch 1 is possible but incredibly ill advised (due to incredibly noisy gradients, please use BS 2 at least) because it does not touch the original model weights.

Fine tuning should only be considered if you have a 24GB vram card and tens of thousands of images.

If you want to try and do this with batch 1, you can try adding accumulation steps to even out the updates but this is also not advised because gradient accumulation has inferior results to plain BS and lowers perf. Using accumulation steps will also allow higher learning rates to be used without breaking the model quickly as well. For e…

Replies: 1 comment 4 replies

Comment options

You must be logged in to vote
4 replies
@Z1-Gamer
Comment options

@Calamdor
Comment options

@Z1-Gamer
Comment options

@Calamdor
Comment options

Answer selected by O-J1
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants