Performance with M4 #263
Replies: 1 comment
-
Hi @jsandlerus There are a couple of things to say on this, all which will impact things. I am going to assume you are looking at AllTalk v2 and not the v1.
To give a bit more context, on a TTS generation, that is generating the whole text (not streaming generation) to the point where I types 290 words, running on CPU with no specific acceleration: XTTSPiperThis is on a Windows Machine, which is typically slower than Unix based OS's and its CPU is nowhere near as powerful as a Mac. Streaming generation would obviously be faster as it is handing over the audio for playback before it has completed generating it e.g. lets say as soon as it has generated 5 seconds of the audio, its playing back, while it continues generating the rest of the TTS.
Finally I do not own a Mac myself, so have no specific way to test out the real hardware, both in terms of performance or compatibility with each individual TTS engine or Mac Metal support. I hope that gives you some insight/answers. Thanks |
Beta Was this translation helpful? Give feedback.
-
I want to use this to power the voice of an llm in my website. The idea is to buy the next mac mini m4 and run this in a local server, then make that server public using ngrok and hit the api every time I need to generate the audio voice. Maybe I can send the file as the response of the local api. My question is, how performant do you think alltalk_tts would be by using the m4 chip? How fast do you think it would take to generate a 400 chars audio?
If I pull this off, we can basically reduce the cost of tts for websites dramatically.
Beta Was this translation helpful? Give feedback.
All reactions