Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Hi, about multi-langual support #67

Open
MonolithFoundation opened this issue Jan 9, 2025 · 2 comments
Open

Hi, about multi-langual support #67

MonolithFoundation opened this issue Jan 9, 2025 · 2 comments

Comments

@MonolithFoundation
Copy link

Would try using WavTokenizer for TTS needs support multilangual, for English it should be OK. but for Chinese or Janpanese, what should be done for better adoption?

thanks for this wonderful work!

@jishengpeng
Copy link
Owner

Would try using WavTokenizer for TTS needs support multilangual, for English it should be OK. but for Chinese or Janpanese, what should be done for better adoption?

thanks for this wonderful work!

Thank you very much for your interest. Building upon the wavtokenizer checkpoint, further training on an additional 10,000 to 20,000 hours of Chinese and Japanese data would likely result in significantly improved performance.

@MonolithFoundation
Copy link
Author

Does the training need words level timestamp? 10000 hours needs many audio data, does not very clean (like Emilla dataset) can fill the constraints?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants