Our great sponsors
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
TTS
:robot: :speech_balloon: Deep learning for Text to Speech (Discussion forum: https://discourse.mozilla.org/c/tts) (by mozilla)
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
-
NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
Thanks! Try the --raw-stream option for listening to long texts: https://github.com/rhasspy/larynx#long-texts
For speech-dispatcher, I'd start a Larynx HTTP server and use curl to get audio. I have an undocumented --daemon flag that does something like this.
A nice enhancement for the system is having TTS read out the currently selected text, triggered by a key shortcut.
I tried festival and it too complicated and my version was too to run the better voices model.
Instead I've used this repo to use upgraded flite: https://github.com/kastnerkyle/hmm_tts_build/
I have mapped keyboard shortcuts Win+1 for normal speed, Win+2 for faster and Win+3 for really fast reading speed. I can use it while reading, to enhance my focus. Neat.
I worked with this a bit not that long ago. For cloud services, quality of Google and Azure "neural" voices are tough to beat. Interestingly I experienced significant latency for all of the Azure services regardless of region, configuration, etc. Never dug deep enough to figure out what was going on there. Also of note, Azure will also let you run their implementation on a local container with the usual "contact us" stuff. Not sure of the terms and pricing on that.
For local, Mozilla TTS was best from a quality standpoint but the GPU inference support was a bit dicey and (possibly) not really supported at all.
For more complex and bespoke applications the Nvidia (I know, I know) NeMO toolkit [0] is very powerful but requires more effort than most to get up and running. However, it provides the ability to do very interesting things with additional training and all things speech.
In the Nvidia world there's also their Riva [1] (formerly Jarvis) solution that works with Triton [2] to build out an architecture for extremely performant and high-scale speech applications with things like model management, revision control, deployment, etc.
[0] https://github.com/NVIDIA/NeMo
[1] https://developer.nvidia.com/riva
[2] https://developer.nvidia.com/nvidia-triton-inference-server
Related posts
- AI-genereeritud Politseikroonika
- Making Voices For System Members
- [Discussion] Is there any open-source alternative to voice.ai ? Looking for open-source speech to speech AI
- Voice actor I need died a decade ago. Is there a program which can create text-to-voice with the voice of a specific person through providing the software voice samples to work from?
- Trying to get it working