encodec
descript-audio-codec
encodec | descript-audio-codec | |
---|---|---|
18 | 2 | |
3,185 | 863 | |
2.0% | 0.0% | |
3.9 | 4.5 | |
4 months ago | about 1 month ago | |
Python | Python | |
MIT License | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
encodec
-
TSAC: Low Bitrate Audio Compression
Since Ballard's codec is "AI" based, can you add google's lyrav2 ( https://github.com/google/lyra ) and Facebook's/meta EnCodec ( https://github.com/facebookresearch/encodec ).
Also I don't seem to be able to access your page, so there might be error.
Finally, when doing opus comparison it's good now to denote if it is using Lace or NoLace decoder post processing filters that became available in opus 1.5 (note, this feature need to be enabled at compile time, and defying decode a new API call needs to be made to force higher complexity decoder) . See https://opus-codec.org/demo/opus-1.5/
-
[R] Neural network for audio training sample size
But models rarely work on raw audio. You can also check EnCodec (https://github.com/facebookresearch/encodec) or SoundStream.
- Bark: A transformer based text to audio system
-
[D]: Is voice cloning or natural TTS (like Elevenlabs) possible due to LLMs?
VALL-E from Microsoft is transformer over Encodec code. SPEAR-TTS from Google is basically AudioLM for TTS.
- Why hasn't Meta made LLaMA open source?
- ML Codecs Similar to Encodec by Facebook?
- Wie Österreich die Glasfaser verschlief - ORF Topos
- EnCodec: State-of-the-art deep learning based audio codec
- High Fidelity Neural Audio Compression
- EnCodec: High Fidelity Neural Audio Compression
descript-audio-codec
-
Show HN: Sonauto – a more controllable AI music creator
Hey HN,
My cofounder (four months ago, classmate) and I trained an AI music generation model and after a month of testing we're launching 1.0 today. Ours is interesting because it's a latent diffusion model instead of a language model, which makes it more controllable: https://sonauto.ai/
Others do music generation by training a Vector Quantized Variational Autoencoder like Descript Audio Codec (https://github.com/descriptinc/descript-audio-codec) to turn music into tokens, then training an LLM on those tokens. Instead, we ripped the tokenization part off and replaced it with a normal variational autoencoder bottleneck (along with some other important changes to enable insane compression ratios). This gave us a nice, normally distributed latent space on which to train a diffusion transformer (like Sora). Our diffusion model is also particularly interesting because it is the first audio diffusion model to generate coherent lyrics!
We like diffusion models for music generation because they have some interesting properties that make controlling them easier (so you can make your own music instead of just taking what the machine gives you). For example, we have a rhythm control mode where you can upload your own percussion line or set a BPM. Very soon you'll also be able to generate proper variations of an uploaded or previously generated song (e.g., you could even sing into Voice Memos for a minute and upload that!). @Musicians of HN, try uploading your songs and using Rhythm Control/let us know what you think! Our goal is to enable more of you, not replace you.
For example, we turned this drum line (https://sonauto.ai/songs/uoTKycBghUBv7wA2YfNz) into this full song (https://sonauto.ai/songs/KSK7WM1PJuz1euhq6lS7 skip to 1:05 if inpatient) or this other song I like better (https://sonauto.ai/songs/qkn3KYv0ICT9kjWTmins we accidentally compressed it with AAC instead of Opus which hurt quality, though)
We also like diffusion models because while they're expensive to train, they're cheap to serve. We built our own efficient inference infrastructure instead of using those expensive inference as a service startups that are all the rage. That's why we're making generations on our site FREE and UNLIMITED for as long as possible.
We'd love to answer your questions. Let us know what you think of our first model! https://sonauto.ai/
-
TSAC: Low Bitrate Audio Compression
Another useful model to compare to would be DAC https://github.com/descriptinc/descript-audio-codec
This is the codec that TSAC extended, so it could be a nice comparison to see. I'd also echo Vocos (from sibling comment), it operates on the same Encodec tokens but generally has better reconstruction quality.
What are some alternatives?
bark - 🚀 BARK INFINITY GUI CMD 🎶 Powered Up Bark Text-prompted Generative Audio Model
bark - 🔊 Text-Prompted Generative Audio Model
bark-with-voice-clone - 🔊 Text-prompted Generative Audio Model - With the ability to clone voices
audiolm-pytorch - Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google Research, in Pytorch