Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

Multi-channel Audio Enhancement #140

Open
TankyFranky opened this issue Dec 1, 2022 · 1 comment
Open

Multi-channel Audio Enhancement #140

TankyFranky opened this issue Dec 1, 2022 · 1 comment

Comments

@TankyFranky
Copy link

Hello,

I am wondering what the best approach would be to adapt the denoiser for multi-channel audio. I have a four microphone array that I would like to apply denoiser to as a pre-processing step.

Can the model.chin and model.chout paramters be changed when performing inference on a network that has been trained on only one channel? Will the inference/forward step adapt if the input tensor is multiple channels of audio (all of the same frame size). I have modified the live.py example to perform sequential forward passes (one for each channel), but obviously this tanks the real time performance.

Any advice on applying denoiser to multi-channel audio would be appreciated.

Thanks.

@adiyoss
Copy link
Contributor

adiyoss commented Dec 2, 2022

Hi @TankyFranky,
You can definitely reconfigured the model to get more than one channel as input and output. However, if you are going that way you should train a new model prom scratch.
If you want to use the pre-trained models, so what you did (process each channel independently) would be the best/easiest way. Regarding the real-time constraints, maybe you can process the channels in parallel?

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants