Story2Hallucination
deep-music-visualizer
Our great sponsors
Story2Hallucination | deep-music-visualizer | |
---|---|---|
13 | 11 | |
146 | 377 | |
-0.7% | - | |
0.0 | 0.0 | |
about 3 years ago | 10 months ago | |
Jupyter Notebook | Python | |
GNU General Public License v3.0 only | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Story2Hallucination
-
test
(Added Feb. 5, 2021) Story2Hallucination.ipynb - Colaboratory by bonkerfield. Uses BigGAN to generate images/videos. GitHub.
- Skrev in "Stockholm" i en AI-generator. Fick fram denna. Känns rimligt ändå
-
Some AI tools I've picked up, and tips.
Third which is pretty new to me is the Story2Hallucination. Which takes text, (say your story) and uses Google Deep Sleep to create a visual using AI to generate images for what the story is describing. And example can be found here.
-
Text to Image Generation
I ran some Lovecraft through Story2Hallucination[1] which uses Big Sleep to make videos from text.
The results were quite something - https://m.imgur.com/tfWLsSR
[1] https://github.com/lots-of-things/Story2Hallucination
-
[P] Visualizing evolution of Text-to-Image generation algorithms side by side by generating video from song-lyrics (X-LXMERT v/s AlpehImage/Dall-E)
Using Story2Hallucination: https://boredhumans.com/music_videos/Story2Hallucination_withwords.mp4 (made with https://github.com/lots-of-things/Story2Hallucination). The problem with it was that I had no easy way to match the timing of the lyrics on the screen with the real singing. So I then made a new version at https://boredhumans.com/music_videos/Story2Hallucination_nowords.mp4 where I edited the code so it did not show the words at all. But it is somewhat boring to watch.
-
Story2Hallucination renders a world of dragons from an AI dungeon game
Here you go: https://github.com/lots-of-things/Story2Hallucination/
-
[D] Will machine learning enable a single person to make a blockbuster movie like Avengers: endgame within 6 months?
This is basic attempt at that: https://github.com/lots-of-things/Story2Hallucination . It turns the text you write into a dream-like series of images converted into an animated GIF. It does not look real, but it is a start.
-
Story2Hallucination render of my latest meme (check comments for links)
Story2Hallucination github by u/bonkerfield : GITHUB LINK
-
Story2Hallucination on another AI Dungeon game and now I have a script to visualize to GIF while playing at the same time.
I've added a slightly simpler notebook to Story2Hallucination that can render GIFs on Google Colab. Note that the story text has to be fairly short or it will make a gigantic unrenderable GIF.
-
Short video generated by Story2Hallucination
Credit to https://github.com/lots-of-things/Story2Hallucination/ and onwards - I converted the script to run in regular python, and on windows through WSL+CUDA (Though, the windows tweaks seem to have caused other issues, Will probably have to roll back and dualboot on this device to do more)
deep-music-visualizer
-
Testing out beat detection
Got inspired by this and started to tinker with librosa. Basically it just zooms and translates on beat.
-
[D] Connecting musical and visual latent spaces in a "harmonic way"
I messed around with this "deep-music-visualizer" package a few years back. I think the concept is close to what you're talking about. It takes an MP3 as input and uses BigGAN to generate images from ImageNet classes, with the rate of change shifting with tempo and class selection with pitch.
-
AI created video for 2nd half of I Know the End
My first attempt at using deep-music-visualizer, so what other song could I have picked ;-)
- The BigGAN model was trained at Google with the cost of approx. 60 000$. Using the model, Deep Music Visualizer generates imagery according to changing pitch and gain of the audio input.
-
I visualize my music using Deep Music Visualizer, which leverages the massive BigGAN model to generate pitch and volume affected imagery.
Software: https://github.com/msieg/deep-music-visualizer
-
Otaku Boy - S3RL ft Jukebox AI
Animation created using: https://github.com/HighCWu/anime_biggan_toy https://github.com/msieg/deep-music-visualizer
-
running "deep music visualizer" in Python
Hi, I'm trying to figure this out to make visuals for my music, and am totally new to Python. I've basically gotten to the point that I can at least confirm I have Python 3 on my Mac in terminal. I am trying to to run this: https://github.com/msieg/deep-music-visualizer
-
Deep Music Visualizer (No coding experience)
I'm a film major and I want to learn how to use Deep Music Visualizer (https://github.com/msieg/deep-music-visualizer) I installed Python 3.9.5 but I dont understand how to even run the visualizer. If anyone has used this program before, please let me know if this is familiar
-
[Project] NEW PYTHON PACKAGE: Sync GAN Art to Music with "Lucid Sonic Dreams"! (Link in Comments)
I am wondering how is this different from Deep Music Visualizer?
-
[P] Visualizing evolution of Text-to-Image generation algorithms side by side by generating video from song-lyrics (X-LXMERT v/s AlpehImage/Dall-E)
Using Deep Music Visualizer - https://boredhumans.com/music_videos/Big_GAN.mp4 (I made it small as a fast demo but it could easily be normal sized) . Made with https://github.com/msieg/deep-music-visualizer. Although the output is good, it uses 12 random classes (or you can pick them manually) of items from ImageNet, and the 1000 classes ImageNet uses really suck for making music videos. They don't have general categories such as beach, ocean, stars. Instead it is very specific things like moving van, drilling platform, and envelope.
What are some alternatives?
deep-daze - Simple command line tool for text to image generation using OpenAI's CLIP and Siren (Implicit neural representation network). Technique was originally created by https://twitter.com/advadnoun
anime_biggan_toy - Generate Amazing Anime Pictures With BigGAN. Just Have Fun !!!
big-sleep - A simple command line tool for text to image generation, using OpenAI's CLIP and a BigGAN. Technique was originally created by https://twitter.com/advadnoun
pretrained-gan-minecraft - Minecraft GAN
DALLE-pytorch - Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
stable-zoomy
StyleCLIP - Official Implementation for "StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery" (ICCV 2021 Oral)
stylized-neural-painting - Official Pytorch implementation of the preprint paper "Stylized Neural Painting", in CVPR 2021.
Voice-Cloning-App - A Python/Pytorch app for easily synthesising human voices
Colab-BigGANxCLIP
stylegan2-clip-approach - Navigating StyleGAN2 w latent space using CLIP