You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello,
We want to perform two-stage fine-tuning with LoRa (SFT, then DPO). We have the first LoRA adapter after SFT, and we want to obtain a second adapter after DPO. Is there a way to continue training the same adapter for DPO using swift, without merging the original model weights with the adapter after SFT?
The text was updated successfully, but these errors were encountered:
Hello,
We want to perform two-stage fine-tuning with LoRa (SFT, then DPO). We have the first LoRA adapter after SFT, and we want to obtain a second adapter after DPO. Is there a way to continue training the same adapter for DPO using
swift
, without merging the original model weights with the adapter after SFT?The text was updated successfully, but these errors were encountered: