Search Results for author: Hiroki Kanagawa

Found 3 papers, 0 papers with code

Noise-robust zero-shot text-to-speech synthesis conditioned on self-supervised speech-representation model with adapters

no code implementations10 Jan 2024 Kenichi Fujita, Hiroshi Sato, Takanori Ashihara, Hiroki Kanagawa, Marc Delcroix, Takafumi Moriya, Yusuke Ijima

The zero-shot text-to-speech (TTS) method, based on speaker embeddings extracted from reference speech using self-supervised learning (SSL) speech representations, can reproduce speaker characteristics very accurately.

Self-Supervised Learning Speech Enhancement +2

Zero-shot text-to-speech synthesis conditioned using self-supervised speech representation model

no code implementations24 Apr 2023 Kenichi Fujita, Takanori Ashihara, Hiroki Kanagawa, Takafumi Moriya, Yusuke Ijima

This paper proposes a zero-shot text-to-speech (TTS) conditioned by a self-supervised speech-representation model acquired through self-supervised learning (SSL).

Self-Supervised Learning Speech Synthesis +1

SIMD-size aware weight regularization for fast neural vocoding on CPU

no code implementations2 Nov 2022 Hiroki Kanagawa, Yusuke Ijima

Pruning time-consuming DNN modules is a promising way to realize a real-time vocoder on a CPU (e. g. WaveRNN, LPCNet).

Cannot find the paper you are looking for? You can Submit a new open access paper.