Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Other Language doesn't generate (zh-cn, ja) #15

Open
MaiChaMH opened this issue May 2, 2024 · 3 comments
Open

Other Language doesn't generate (zh-cn, ja) #15

MaiChaMH opened this issue May 2, 2024 · 3 comments

Comments

@MaiChaMH
Copy link

MaiChaMH commented May 2, 2024

I'm using the 0.0.7 version with xtts (installed through one click exe). English and some other languages seem to be doing fine. However, when I select another language (zh-cn & ja specifically) it doesn't generate.

There's nothing happening after pressing the Start generation button, and to confirm I clicked the cancel generation button after the start button it says that no process is happening (also went to task manager to double check, no CPU process).

@lukaszliniewicz
Copy link
Owner

Sorry about that, I think I know what's wrong - the sentence splitting library I'm using does not support either Chinese or Japanese. I have to use a different one for these languages. I will post an update soon.

@lukaszliniewicz
Copy link
Owner

Please try installing it again (remove the Pandrator folder first). Japanese should work now, but I'm not sure if splitting sentences into shorter chunks based on the character limit set under "Text processing" works well or how many Japanese characters XTTS can process well in one go. Please let me know what you find.

@MaiChaMH
Copy link
Author

MaiChaMH commented May 2, 2024

Please try installing it again (remove the Pandrator folder first). Japanese should work now, but I'm not sure if splitting sentences into shorter chunks based on the character limit set under "Text processing" works well or how many Japanese characters XTTS can process well in one go. Please let me know what you find.

I've re-installed it (by removing the folder). Japanese tts generation is working fine now.

In terms of Text Processing, I think it's working fine. I've fed it a txt containing around 4000 Japanese characters, tested with paragraph spacing and without (paragraph spacing works fine). I've tested length 30 and length 3000, with 30 it'd successfully split at the end of a sentence, with 3000 XTTS handled it in one go with no problems and no artifacts in the audio (very powerful).

image

Also for Chinese libraries, I've found (By Asking Bing AI) these 2 libraries that might be helpful. The pywordseg and jieba library, according to the search results from the AI, it seemed that pywordseg is for word segmentation, and jieba can do both word segmentation and sentence splitting. (Both libraries can do simplified and traditional Chinese characters.)

image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants