mimic-recording-studio
community-events
mimic-recording-studio | community-events | |
---|---|---|
4 | 8 | |
486 | 385 | |
0.0% | 2.1% | |
0.0 | 7.2 | |
about 1 year ago | 6 months ago | |
JavaScript | Jupyter Notebook | |
Apache License 2.0 | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
mimic-recording-studio
-
How can I create a dataset to refine Whisper AI from old videos with subtitles?
I weirdly can't find a great off-the-shelf app for this. l'd love to know if anyone finds one. Most stuff seems to be for recording data for Text To Speech (going the other way). Mimic Recording Studio looks the best. Then there's speech training recorder and TTS Dataset Creator (video). You don't have to worry about audio quality as much as they do.
-
Experts warn people not to put Amazon Echo Alexa devices in their bedrooms
If you genuinely want to help out in that way, you can create a voice for Mycroft using their Mimic recording studio software. They reckon it takes about 40-80 hours all up.
-
Mimic 3 by Mycroft
They want you to make good quality audio recordings of you speaking about 20 000 phrases. It could take 40 to 80 hours of speaking and recording, maximum 4 hours per day.
https://github.com/MycroftAI/mimic-recording-studio
https://mycroft.ai/contribute/
- How can I immortalise my dying mums voice in some sort of voice assistant?
community-events
-
Controlling Stable Diffusion with JAX & Diffusers using TPU v4
Best applications that will come out of this sprint will receive prizes. You can find more information here. If you want to get started, simply join huggingface.co/discord, take the role 🧨 Diffusers and head to #jax-diffusers-ideas to share your idea or join one of the teams, and fill this form: https://forms.gle/t3M7aNPuLL9V1sfa9
-
JAX & Diffusers to Control Stable Diffusion (with TPUs ⚡️ )
It will start on 17th of April. To join us, you can join huggingface.co/join/discord and take the role Diffusers from #role-assignment. After this, simply fill the form provided in this guide to later get access to TPUs. https://github.com/huggingface/community-events/tree/main/jax-controlnet-sprint
- “Control Stable Diffusion” Sprint kicks off with free TPU-v4 from Google
-
Free compute to train custom ControlNet by Hugging Face
Details and sign-up: https://github.com/huggingface/community-events/tree/main/jax-controlnet-sprint
-
How can I create a dataset to refine Whisper AI from old videos with subtitles?
For the training, I extremely recommend checking out the Whisper Fine-Tuning Event. It has a python script to train in one command, tons of tips, even a walkthrough video.
- I am using OpenAi's whisper transcription/translation model. I am wondering if I can improve it's performance by optimizing the audio files somehow. What features of audio files should I look into to make the whisper model perform better?
-
[N] Gradio Blocks + Hugging Face event is starting this week. A hackathon type event from May 17th to May 31st with prizes in which we will create interactive web demos for state-of-the-art machine learning models
We are happy to invite you to the Gradio Blocks Party - a community event in which we will create interactive demos for state-of-the-art machine learning models. Demos are powerful because they allow anyone — not just ML engineers — to try out models in the browser, give feedback on predictions, identify trustworthy models. The event will take place from May 17th to 31st. We will be organizing this event on Github and the Hugging Face discord channel. Prizes will be given at the end of the event, see the Prizes section
-
Dall-E 2
If you're interested in generative models, Hugging Face is putting on an event around generative models right now called the HugGAN sprint, where they're giving away free access to compute to train models like this.
You can join it by following the steps in the guide here: https://github.com/huggingface/community-events/tree/main/hu...
There will also be talks from awesome folks at EleutherAI, Google, and Deepmind
What are some alternatives?
Real-Time-Voice-Cloning - Clone a voice in 5 seconds to generate arbitrary speech in real-time
dalle-2-preview
annyang - :speech_balloon: Speech recognition for your site
dalle-mini - DALL·E Mini - Generate images from a text prompt
google-tts - Google TTS (Text-To-Speech) for node.js
bevy_retro - Plugin pack for making 2D games with Bevy
speaker.app - Speaker.app (https://speaker.app) is an encrypted peer-to-peer (P2P) group communication platform which does not require a user account to use. Users remain anonymous on the network unless choosing to personally identify themselves.
lm-human-preferences - Code for the paper Fine-Tuning Language Models from Human Preferences
mimic3 - A fast local neural text to speech engine for Mycroft
gpt-3 - GPT-3: Language Models are Few-Shot Learners
Portainer - Making Docker and Kubernetes management easy.
glide-text2im - GLIDE: a diffusion-based text-conditional image synthesis model