← Back to context

Comment by hosaka

12 hours ago

Depending on the TTS model being used latency can be reduced further yet with an LRU cache, fetching common phrases from cache instead of generating fresh with TTS.

However the naturalness of how it sounds will depend on how the TTS model works and whether two identical chunks of text will sound alike every generation.