This is a demonstration page for the following paper.
Tsubasa Ochiai, Marc Delcroix, Yuma Koizumi, Hiroaki Ito, Keisuke Kinoshita, Shoko Araki,
``Listen to What You Want:
Neural Network-based Universal Sound Selector'', submitted to Interspeech 2020.
Here are a set of audio samples for the proposed universal sound selector/remover:
Waveform (a) corresponds to observed signal, (b)(c) reference signal for sound selection task, (d)(e) reference signal for sound removal task,
(b')(c') estimated signal by unversal sound selector and (d')(e') estimated signal by universal sound remover.