Llama-3.3-70B-Instruct-4bit LoRA Fine-Tuning: No Change (or Instability) - Adapter Issue? #1147
-
Hi everyone, The core problem is that the LoRA adapter seems to be having no usable effect on the model's output, despite successful training (loss decreases normally). It's not a matter of fine-tuning the scale - it's like the adaptation either does nothing or breaks the model. Here's what I've tried:
I'm really stuck here, and any insights or suggestions would be greatly appreciated! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
I tried training this:
And then evaluating it like this:
And it generated the following which is very reasonable:
So I'm not sure where things are going wrong for you. A few suggestions:
|
Beta Was this translation helpful? Give feedback.
I tried training this:
And then evaluating it like this:
And it generated the following which is very reasonable: