Bottleneck Configs do not work with ln_before = True
and init_weights = "mam_adapter"
#745
Labels
bug
Something isn't working
Given that we specify a bottleneck config with layer normalization before the adapter bottle neck, and we specify the initial weights of the adapter to be of type
mam_adapter
, it will trigger an error.Reproducible Code
This is because if we set$W_{down}$ but it is actually a normalization layer.
ln_before
= True, the first layer ofseq_list
for the will be a normalization layer. Then when we attempt to initialize the weights using the typemam_adapter
, the kaiming uniform will be applied onto the first layer of theseq_list
thinking it isThe text was updated successfully, but these errors were encountered: