You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I noticed that you apply the mul operation in LoraA and LoraB, then, you sum the result with the input.
I think the result of multiplying LoraA and LoraB has to be summed to the original weights, or I am wrong?
scaling follows the original implementation https://github.com/microsoft/LoRA
It's mentioned in the paper. From my understanding it's not important, it's only there to control for the change of rank.
I noticed that you apply the mul operation in LoraA and LoraB, then, you sum the result with the input.
I think the result of multiplying LoraA and LoraB has to be summed to the original weights, or I am wrong?
Could you also explain the scaling factor?
Thanks.
The text was updated successfully, but these errors were encountered: