Reproducible gradient norm spike in QLoRA at step 44 on Mistral-7B — spectral norm constraint as a fix #3057
fourwheels2512
started this conversation in
General
Replies: 1 comment
-
|
Gradient norm spikes are not too unusual, especially when dealing with lower quantizations (as steps get more discrete). Have you tested using |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
While testing several PEFT techniques on Mistral-7B, we noticed a reproducible gradient norm spike specifically with QLoRA at training step 44.
What we observed:
Why AdaLoRA and VeRA don't fully solve this:
What worked for us:
Adding a spectral norm constraint on top of the QLoRA training loop. We built this into a free tool (no GPU required) that you can use to fine-tune and compare results:
https://huggingface.co/spaces/Fourwheels2512/crma-fine-tuner
Happy to share more details on the spectral norm implementation if useful. Has anyone else hit this step-44 spike or similar instability patterns with QLoRA on 7B-class models?
Beta Was this translation helpful? Give feedback.
All reactions