iSTFTNet2

Faster and More Lightweight iSTFT-Based Neural Vocoder Using 1D-2D CNN

Interspeech 2023

Paper

iSTFTNet2
Figure 1. Comparison of iSTFTNet [1] and iSTFTNet2 (proposed). (a) Owing to the difficulty of a 1D CNN to model high-dimensional spectrograms, it is necessary to reduce the frequency dimension sufficiently in iSTFTNet using large temporal upsampling. (b) In contrast, iSTFTNet2 mitigates this difficulty by conducting 1D-to-2D conversion in an earlier stage and applying a 2D CNN that can capture local structures in spectrograms. This modification facilitates the reduction of the neural temporal upsampling by eight times (i.e., from \( \times 64 \) to \( \times 8 \)) and enhances the inference speed.

Note

Check out our relevant work: iSTFTNet / MISRNet / Wave-U-Net Discriminator / AugCondD

Abstract

The inverse short-time Fourier transform network (iSTFTNet) [1] has garnered attention owing to its fast, lightweight, and high-fidelity speech synthesis. It obtains these characteristics using a fast and lightweight 1D CNN (e.g., HiFi-GAN [2]) as the backbone and replacing some neural processes with iSTFT. Owing to the difficulty of a 1D CNN to model high-dimensional spectrograms, the frequency dimension is reduced via temporal upsampling. However, this strategy compromises the potential to enhance the speed. Therefore, we propose iSTFTNet2, an improved variant of iSTFTNet with a 1D-2D CNN that employs 1D and 2D CNNs to model temporal and spectrogram structures, respectively. We designed a 2D CNN that performs frequency upsampling after conversion in a few-frequency space. This design facilitates the modeling of high-dimensional spectrograms without compromising the speed. The results demonstrated that iSTFTNet2 made iSTFTNet faster and more lightweight with comparable speech quality.

Contents

  1. Comparison on LJSpeech
  2. Comparison on VCTK

Results

I. Comparison on LJSpeech
Sample 1 Sample 2 Sample 3 Sample 4 Sample 5 Sample 6
Ground truth
HiFi-GAN V2 [2]
iSTFTNet-C8C8I4 [1]
iSTFTNet-C8C1I32 [1]
iSTFTNet2-Base
iSTFTNet2-Small
iSTFTNet-MB [1][4]
iSTFTNet2-MB
II. Comparison on VCTK
  • Dataset: VCTK [5] (multiple English speakers)
Sample 1 (p227) Sample 2 (p236) Sample 3 (p245) Sample 4 (p272) Sample 5 (p312) Sample 6 (p339)
Ground truth
HiFi-GAN V2 [2]
iSTFTNet-C8C8I4 [1]
iSTFTNet-C8C1I32 [1]
iSTFTNet2-Base
iSTFTNet2-Small

Citation

@inproceedings{kaneko2023istftnet2,
  title={{iSTFTNet2}: Faster and More Lightweight {iSTFT}-Based Neural Vocoder Using {1D-2D CNN}},
  author={Takuhiro Kaneko and Hirokazu Kameoka and Kou Tanaka and Shogo Seki},
  booktitle={Interspeech},
  year={2023},
}

References

  1. T. Kaneko, K. Tanaka, H. Kameoka, S. Seki. iSTFTNet: Fast and Lightweight Mel-Spectrogram Vocoder Incorporating Inverse Short-Time Fourier Transform. ICASSP, 2022.
  2. J. Kong, J. Kim, J. Bae. HiFi-GAN: Generative Adversarial Networks for Efficient and High Fidelity Speech Synthesis. NeurIPS, 2020.
  3. K. Ito, L. Johnson. The LJ Speech Dataset. 2017.
  4. G. Yang, S. Yang, K. Liu, P. Fang, W. Chen, L. Xie. Multi-Band Melgan: Faster Waveform Generation For High-Quality Text-To-Speech. SLT, 2021.
  5. J. Yamagishi, C. Veaux, K. MacDonald. CSTR VCTK Corpus: English Multi-speaker Corpus for CSTR Voice Cloning Toolkit. University of Edinburgh. The Centre for Speech Technology Research, 2017.