-
I need an output that should maintain some of the content input style. Is there a way to conserve input music like 40 ~ 50 percent and add a different style? Maybe this is the limitation of the current model? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Normally only the harmony is preserved from the content input. Neither style nor the melody of the content input is preserved. This is why I'm adding the content melody to the output. To achieve what you want, you could try to interpolate the styles of both songs as described in section VI.E of the paper. You would need to feed both songs to the style encoder to obtain two style embeddings |
Beta Was this translation helpful? Give feedback.
Normally only the harmony is preserved from the content input. Neither style nor the melody of the content input is preserved. This is why I'm adding the content melody to the output.
To achieve what you want, you could try to interpolate the styles of both songs as described in section VI.E of the paper. You would need to feed both songs to the style encoder to obtain two style embeddings
s1
ands2
and then "blend" them e.g. withs = 0.4 * s1 + 0.6 * s2
and feed thiss
to the decoder as the new style embedding. But note that this still won't allow you to preserve any specific parts of the content input, e.g. there will still be no melody. All the model will try to do is come up with an a…