audio-diffusion
Apply diffusion models using the new Hugging Face diffusers package to synthesize music instead of images. (by teticio)
riffusion-inference
Stable diffusion for real-time music generation [Moved to: https://github.com/riffusion/riffusion-inference] (by hmartiro)
audio-diffusion | riffusion-inference | |
---|---|---|
1 | 6 | |
653 | 714 | |
- | - | |
5.0 | 10.0 | |
6 months ago | over 1 year ago | |
Jupyter Notebook | Python | |
GNU General Public License v3.0 only | MIT License |
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
audio-diffusion
Posts with mentions or reviews of audio-diffusion.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2022-12-15.
riffusion-inference
Posts with mentions or reviews of riffusion-inference.
We have used some of these posts to build our list of alternatives
and similar projects. The last one was on 2022-12-16.
- Looping/Interpolation
-
Riffusion Manipulation Tools
Just put the generated image into the seeds folder on the inference server, replace it with one of them on the seed_images folder: https://github.com/hmartiro/riffusion-inference/tree/main/seed_images
-
Am I understanding this right?
I think so. This is the specific code they say they use to do the image>audio conversion
-
Stable Diffusion fine-tuned to generate Music — Riffusion
I'm still reading, but it looks like they're doing some extra pre and post processing: https://github.com/hmartiro/riffusion-inference
-
Riffusion – Stable Diffusion fine-tuned to generate Music
Yes from https://huggingface.co/runwayml/stable-diffusion-v1-5. Our checkpoint works with automatic1111, and if you'd like to make an extension to decode to audio, it should be pretty straightforward: https://github.com/hmartiro/riffusion-inference/blob/main/ri...
What are some alternatives?
When comparing audio-diffusion and riffusion-inference you can also consider the following projects:
bumblebee - Pre-trained Neural Network models in Axon (+ 🤗 Models integration)
sd-webui-riffusion - Riffusion extension for AUTOMATIC1111's SD Web UI