-
stylegan2-pytorch
Simplest working implementation of Stylegan2, state of the art generative adversarial network, in Pytorch. Enabling everyone to experience disentanglement
-
CogVideo
Text-to-video generation. The repo for ICLR2023 paper "CogVideo: Large-scale Pretraining for Text-to-Video Generation via Transformers"
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Seeing as there's been a recent onslaught of very negative sentiment in the past few days, with a lot of people feeling very afraid (Lots of digital artists worried about their future; AI artists afraid that the wave of Anti-AI sentiment will lead to their tools to get banned), I thought I'd write something perhaps a bit more positive. I'd been following media synthesis since the old days of deep dream and StyleGAN that could create lots of trippy eyes, human faces and fursonas, following the development of transformer models and the days of blurry Dall-E 1 outputs down to Dall-E 2, Disco Diffusion and Stable Diffusion.
The second thing that will happen is the appearance of AI video and audio. Google has shown two programs for video generation, one which is fairly high quality and the other which can make long videos with several scenes. Meta has also demonstrated their own. We've already seen other projects like CogVideo, as well as many others that are currently being worked on. It's likely that these techniques will become so refined that over the next year or two, they'll have a similar boom to image generation programs. And eventually, they'll have a similar application in video editing, once coherence is adequate enough. Select a person's shirt, and it stays that for the remainder of the scene. Change an actor's hairstyle in real time, or add characters that didn't exist into a scene and let the computer figure out the desired level of realism. This'll revolutionize VFX to a degree where making an effects heavy will be less about wrangling complex toolsets and more about making aesthetic choices of style and placement.