-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Rendering in Stereo? #131
Comments
One way to make fake stereo, during inference: you could make two upsampled versions of the final tier. They'll sound slightly different, it'll probably be cool. Your bass should be mono though, to prevent phase cancelation and mud, so mix everything below ~200Hz as mono (use only one version of the audio, don't combine the channels into mono for bass) |
Great idea—it worked! I had to re-run tiers 0 & 1 instead of just 0, but that provided about the right amount of variation. Maybe even a third run for the mono bass would sound good. Another cool way to get around the stereo problem is to feed Jukebox a loop as a primer and then send the sparser continuations left and right in a DAW. |
Wait, how can it be possible to get the stereo effect this way? The two channels are supposed to provide spatial positioning information, but if you just generate two slightly different variations of the same things it would be nothing of the sort. Can you please post an example of what you made? |
Great idea—it worked! I had to re-run tiers 0 & 1 instead of just 0, but that provided about the right amount of variation. Maybe even a third run for the mono bass would sound good. Another cool way to get around the stereo problem is to feed Jukebox a loop as a primer and then send the sparser continuations left and right in a DAW. |
It’s fake stereo, but I like that it’s not what you’d hear in the real world. At some point Jukebox will be stereo and near-perfect, so I am taking advantage of its current flaws. I don’t want to post an example at this point but you can try it to see if it works for you. |
Another technique would be to split the stems of your generated tracks
using Spleeter or OpenUnmix or Wave-U-Net, and then widen/pan them around
in a DAW
…On Tue, Aug 18, 2020 at 4:37 PM Demon Flex Council ***@***.***> wrote:
It’s fake stereo, but I like that it’s not what you’d hear in the real
world. At some point Jukebox will be stereo and near-perfect, so I am
taking advantage of its current flaws. I don’t want to post an example at
this point but you can try it to see if it works for you.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#131 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/AASXCZID2OJJXAG5GSZID63SBLRCHANCNFSM4PQXXETA>
.
|
Anybody know how to make stereo renders with Jukebox? I tried changing a hyperparameter (hps.channels = 2) but the model was expecting a tensor with “1” in the second dimension, implying mono:
AssertionError: Expected (1, 1, 831872) got torch.Size([1, 2, 831872])
Full traceback:
It seemed like the problem was with the variable “input_emb_width”, which is hardcoded in make_vqvae to “1”. I tried changing that to “2” but ran into this error:
Is mono just baked into the training? Since training is cost-prohibitive, is there any way around this at the inference stage? Jukebox would sound so much richer in stereo.
The text was updated successfully, but these errors were encountered: