论文标题

用于模块化源分离的潜在迭代精致

Latent Iterative Refinement for Modular Source Separation

论文作者

Bralios, Dimitrios, Tzinis, Efthymios, Wichern, Gordon, Smaragdis, Paris, Roux, Jonathan Le

论文摘要

传统的源分离方法端到端训练深度神经网络模型,并通过最大程度地降低整个训练集的经验风险,并立即获得所有可用的数据。在推理方面,训练模型后,用户获取静态计算图,并在某些指定的观察到的混合信号上运行完整模型以获取估计的源信号。此外,其中许多模型由几个基本的处理块组成,这些块被顺序应用。我们认为,通过将模型的培训和推理程序重新提高作为潜在信号表示的迭代映射,我们可以显着提高培训和推理阶段的资源效率。首先,我们可以在其输出上多次应用相同的处理块,以完善输入信号,从而提高参数效率。在培训期间,我们可以遵循一个宽阔的过程,从而可以减少内存需求。因此,与端到端培训相比,可以使用计算明显较少的计算来训练非常复杂的网络结构。在推断期间,我们可以动态调整使用门控模块的输入信号需求的特定块的处理块和迭代。

Traditional source separation approaches train deep neural network models end-to-end with all the data available at once by minimizing the empirical risk on the whole training set. On the inference side, after training the model, the user fetches a static computation graph and runs the full model on some specified observed mixture signal to get the estimated source signals. Additionally, many of those models consist of several basic processing blocks which are applied sequentially. We argue that we can significantly increase resource efficiency during both training and inference stages by reformulating a model's training and inference procedures as iterative mappings of latent signal representations. First, we can apply the same processing block more than once on its output to refine the input signal and consequently improve parameter efficiency. During training, we can follow a block-wise procedure which enables a reduction on memory requirements. Thus, one can train a very complicated network structure using significantly less computation compared to end-to-end training. During inference, we can dynamically adjust how many processing blocks and iterations of a specific block an input signal needs using a gating module.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源