Speech-to-Speech Translation Between Untranscribed Unknown Languages
2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU)2019pp. 593–600
Citations Over TimeTop 10% of 2019 papers
Abstract
In this paper, we explore a method for training speech-to-speech translation tasks without any transcription or linguistic supervision. Our proposed method consists of two steps: First, we train and generate discrete representation with unsupervised term discovery with a discrete quantized autoencoder. Second, we train a sequence-to-sequence model that directly maps the source language speech to the target languages discrete representation. Our proposed method can directly generate target speech without any auxiliary or pre-training steps with a source or target transcription. To the best of our knowledge, this is the first work that performed pure speech-to-speech translation between untranscribed unknown languages.
Related Papers
- → Performance Comparison of Three Types of Autoencoder Neural Networks(2008)28 cited
- → The Learning Effect of Different Hidden Layers Stacked Autoencoder(2016)20 cited
- → Combining an Autoencoder and a Variational Autoencoder for Explaining the Machine Learning Model Predictions(2021)5 cited
- → Autoencoder: An Unsupervised Deep Learning Approach(2022)4 cited
- → A Comparative Evaluation of AutoEncoder-Based Unsupervised Anomaly Detection Methods Applied on Space Payload(2020)