Possible fix of wrong scale in weight decomposition#16151
Conversation
|
Hi, this patch is wrong. This variable is used by model authors to normalize the trained LoRA/DoRA from -1 to 1. With this patch the DoRA breaks when alpha changes instead of normalizing. |
|
it breaks bcuz it is not how alpha works. |
|
What is this useful for? |
wrong formula: W + scale * alpha/dim * (wd(W + BA) - W) I think this is very intuitive that alpha/dim should not affect the "- W" part. |
|
Also that's how training of DoRA in my repo works. |
|
Hi, this PR should be reverted. Here are the reasons why:
|
|
I hope you can understand the impact of these changes on users. If you have the 'special weight scaling' code, feel free to share it. I don't think its possible in this formulation. |
Description
Should resolve this: Comfy-Org/ComfyUI#3922
Checklist: