bnew

Veteran
Joined
Nov 1, 2015
Messages
51,805
Reputation
7,926
Daps
148,748

jLXBeG2.png
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,805
Reputation
7,926
Daps
148,748

As an AI language model, I do not have personal experience or habits. However, here are some ways in which you can use LLM (Large Language Models) in your daily routine:
1Writing and editing documents: LLM models can help you improve the quality of your writing by providing suggestions for grammar, punctuation, and phrasing. You can also use them to edit and proofread your documents more efficiently.
2Translation: If you work with people from different parts of the world, LLM models can help you translate your messages into different languages more accurately and quickly.
3Speech recognition: You can use LLM models to transcribe your speech and turn it into text. This can be useful when you need to take notes during a meeting or lecture.
4Research: LLM models can help you gather information on various topics by summarizing articles and providing insights. This can be useful if you need to research a topic for a presentation or report.
5Personal assistant: LLM models can act as your personal assistant by scheduling appointments, reminding you of important dates, and answering your questions.
Overall, LLM models can save you time and help you communicate more effectively in your daily routine
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,805
Reputation
7,926
Daps
148,748



Demo of reproduced VALL-E X​

Original description from Microsoft​

VALL-E X can synthesize personalized speech in another language for a monolingual speaker. Taking the phoneme sequences derived from the source and target text, and the source acoustic tokens derived from an audio codec model as prompts, VALL-E X is able to produce the acoustic tokens in the target language, which can be then decompressed to the target speech waveform. Thanks to its powerful in-context learning capabilities, VALL-E X does not require cross-lingual speech data of the same speakers for training and can perform various zero-shot cross-lingual speech generation tasks, such as cross-lingual text-to-speech synthesis and speech-to-speech translation.

Additional description for reproduced model​

Data used for trainingEnglishChineseJapanese
Microsoft'sLibriLight (70k+ hours)Wenet Speech (10k+ hours)-
Ours (reproduced)LibriTTS + self-gathered (704 hours)Aishell 1, 3, Aidatatang + self-gathered (598 hours)JP commonvoice + self-gathered (437 hours)
Unlike Microsoft's original model, the reproduced version does not contain the Speech-to-speech translation (S2ST) module (it requires additional training), but this functionality can be achieved by using Google Translate API.
VALL-E paper: Neural Codec Language Models are Zero-Shot Text to Speech Synthesizers

VALL-E X paper: Speak Foreign Languages with Your Own Voice: Cross-Lingual Neural Codec Language Modeling

official demo page: VALL-E (X)

UI & API for usage: GitHub - Plachtaa/VALL-E-X: An open source implementation of Microsoft's VALL-E X zero-shot TTS model. Demo is available in https://plachtaa.github.io

Training code from: GitHub - lifeiteng/vall-e: PyTorch implementation of VALL-E(Zero-Shot Text-To-Speech), Reproduced Demo https://lifeiteng.github.io/valle/index.html

This page is for showing reproduced results only.

Model Overview​

vallex_framework.jpg

VALL-E X can synthesize personalized speech in another language for a monolingual speaker. Taking the phoneme sequences derived from the source and target text, and the source acoustic tokens derived from an audio codec model as prompts, VALL-E X is able to produce the acoustic tokens in the target language, which can be then decompressed to the target speech waveform. Thanks to its powerful in-context learning capabilities, VALL-E X does not require cross-lingual speech data of the same speakers for training and can perform various zero-shot cross-lingual speech generation tasks, such as cross-lingual text-to-speech synthesis and speech-to-speech translation.
 
Top