Search Results for author: Tamás Gábor Csapó

Found 12 papers, 8 papers with code

Towards Decoding Brain Activity During Passive Listening of Speech

1 code implementation26 Feb 2024 Milán András Fodor, Tamás Gábor Csapó, Frigyes Viktor Arthur

The aim of the study is to investigate the complex mechanisms of speech perception and ultimately decode the electrical changes in the brain accruing while listening to speech.

Brain Computer Interface Speech Synthesis

Adaptation of Tacotron2-based Text-To-Speech for Articulatory-to-Acoustic Mapping using Ultrasound Tongue Imaging

1 code implementation26 Jul 2021 Csaba Zainkó, László Tóth, Amin Honarmandi Shandiz, Gábor Gosztolya, Alexandra Markó, Géza Németh, Tamás Gábor Csapó

In this paper, we experimented with transfer learning and adaptation of a Tacotron2 text-to-speech model to improve the final synthesis quality of ultrasound-based articulatory-to-acoustic mapping with a limited database.

Transfer Learning

Extending Text-to-Speech Synthesis with Articulatory Movement Prediction using Ultrasound Tongue Imaging

1 code implementation12 Jul 2021 Tamás Gábor Csapó

In this paper, we present our first experiments in text-to-articulation prediction, using ultrasound tongue image targets.

Speech Synthesis Text-To-Speech Synthesis

Speech Synthesis from Text and Ultrasound Tongue Image-based Articulatory Input

1 code implementation5 Jul 2021 Tamás Gábor Csapó, László Tóth, Gábor Gosztolya, Alexandra Markó

Besides, we analyze the ultrasound tongue recordings of several speakers, and show that misalignments in the ultrasound transducer positioning can have a negative effect on the final synthesis performance.

Speech Synthesis Text-To-Speech Synthesis

Advances in Speech Vocoding for Text-to-Speech with Continuous Parameters

no code implementations19 Jun 2021 Mohammed Salah Al-Radhi, Tamás Gábor Csapó, Géza Németh

Vocoders received renewed attention as main components in statistical parametric text-to-speech (TTS) synthesis and speech transformation systems.

Speech Synthesis

Continuous Wavelet Vocoder-based Decomposition of Parametric Speech Waveform Synthesis

no code implementations12 Jun 2021 Mohammed Salah Al-Radhi, Tamás Gábor Csapó, Csaba Zainkó, Géza Németh

To date, various speech technology systems have adopted the vocoder approach, a method for synthesizing speech waveform that shows a major role in the performance of statistical parametric speech synthesis.

Speech Synthesis

Towards a practical lip-to-speech conversion system using deep neural networks and mobile application frontend

no code implementations29 Apr 2021 Frigyes Viktor Arthur, Tamás Gábor Csapó

Articulatory-to-acoustic (forward) mapping is a technique to predict speech using various articulatory acquisition techniques as input (e. g. ultrasound tongue imaging, MRI, lip video).

Lip to Speech Synthesis Speech Synthesis

Quantification of Transducer Misalignment in Ultrasound Tongue Imaging

1 code implementation6 Aug 2020 Tamás Gábor Csapó, Kele Xu

Ultrasound is an affordable and non-invasive imaging modality with relatively high temporal and spatial resolution to study the dynamic behavior of tongue during speech production.

Audio and Speech Processing Sound

Ultrasound-based Articulatory-to-Acoustic Mapping with WaveGlow Speech Synthesis

1 code implementation6 Aug 2020 Tamás Gábor Csapó, Csaba Zainkó, László Tóth, Gábor Gosztolya, Alexandra Markó

The training target is the 80-dimensional mel-spectrogram, which results in a finer detailed spectral representation than the previously used 25-dimensional Mel-Generalized Cepstrum.

Audio and Speech Processing Sound

Speaker dependent acoustic-to-articulatory inversion using real-time MRI of the vocal tract

1 code implementation4 Aug 2020 Tamás Gábor Csapó

Acoustic-to-articulatory inversion (AAI) methods estimate articulatory movements from the acoustic speech signal, which can be useful in several tasks such as speech recognition, synthesis, talking heads and language tutoring.

Audio and Speech Processing Sound

Speaker dependent articulatory-to-acoustic mapping using real-time MRI of the vocal tract

1 code implementation3 Aug 2020 Tamás Gábor Csapó

Articulatory-to-acoustic (forward) mapping is a technique to predict speech using various articulatory acquisition techniques (e. g. ultrasound tongue imaging, lip video).

Ultrasound-based Silent Speech Interface Built on a Continuous Vocoder

no code implementations24 Jun 2019 Tamás Gábor Csapó, Mohammed Salah Al-Radhi, Géza Németh, Gábor Gosztolya, Tamás Grósz, László Tóth, Alexandra Markó

Recently it was shown that within the Silent Speech Interface (SSI) field, the prediction of F0 is possible from Ultrasound Tongue Images (UTI) as the articulatory input, using Deep Neural Networks for articulatory-to-acoustic mapping.

Sound Audio and Speech Processing

Cannot find the paper you are looking for? You can Submit a new open access paper.