Our great sponsors
-
InvokeAI
InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Maybe i misunderstand but with InvokeAI you can load different models, not just a single one.
You can try running GPT 2, I had some fun playing around with it, but it struggles with context. GPT 3 and now 3.5 blows it way out the water, but even if the model was available, you would need at least 350 GB of VRAM to run it. So unfortunately, for now it's out of reach.
The closest thing available at the moment is GPT-J and GPT-NeoX, by ElutherAI. I'm sure they will be doing their best to catch up to OpenAI, if it's possible.
https://github.com/onury5506/Discord-ChatGPT-Bot is what I'm using right now. It uses this: https://github.com/transitive-bullshit/chatgpt-api
https://github.com/onury5506/Discord-ChatGPT-Bot is what I'm using right now. It uses this: https://github.com/transitive-bullshit/chatgpt-api
Such a machine wouldn't be able to run GPT-3. Consider OpenAI Whisper. While it's a different model we can still get some numbers about what it takes to run. The large model for Whisper is 2.87 GB, but requires 10 GB vram to run. Again, it's not apples to apples, but one can assume that it would take significantly more than 24 GB vram to run an 800 GB model.